Abstract
Many applications require learning classifiers or regressors that are both accurate and cheap to evaluate. Prediction cost can be drastically reduced if the learned predictor is constructed such that on the majority of the inputs, it uses cheap features and fast evaluations. The main challenge is to do so with little loss in accuracy. In this work we propose a budget-aware strategy based on deep boosted regression trees. In contrast to previous approaches to learning with cost penalties, our method can grow very deep trees that on average are nonetheless cheap to compute. We evaluate our method on a number of datasets and find that it outperforms the current state of the art by a large margin. Our algorithm is easy to implement and its learning time is comparable to that of the original gradient boosting. Source code is made available at http://github.com/svenpeter42/LightGBM-CEGB.
Original language | English |
---|---|
Title of host publication | Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS 2017) |
Editors | UV Luxburg, I Guyon, S Bengio, H Wallach, R Fergus |
Pages | 1550–1560 |
Number of pages | 11 |
State | Published - Dec 2017 |
Event | 31st Conference on Neural Information Processing Systems - Long Beach Convention Center, Long Beach, United States Duration: 4 Dec 2017 → 9 Dec 2017 Conference number: 31st |
Publication series
Name | Advances in Neural Information Processing Systems |
---|---|
Volume | 30 |
ISSN (Print) | 1049-5258 |
Conference
Conference | 31st Conference on Neural Information Processing Systems |
---|---|
Abbreviated title | NIPS'17 |
Country/Territory | United States |
City | Long Beach |
Period | 4/12/17 → 9/12/17 |