2022
DOI: 10.48550/arxiv.2205.11412
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Instance-Based Uncertainty Estimation for Gradient-Boosted Regression Trees

Abstract: Gradient-boosted regression trees (GBRTs) are hugely popular for solving tabular regression problems, but provide no estimate of uncertainty. We propose Instance-Based Uncertainty estimation for Gradient-boosted regression trees (IBUG), a simple method for extending any GBRT point predictor to produce probabilistic predictions. IBUG computes a non-parametric distribution around a prediction using the k-nearest training instances, where distance is measured with a treeensemble kernel. The runtime of IBUG depend… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 26 publications
0
1
0
Order By: Relevance
“…Gradient boosting decision trees are among the most popular ML models, which often outperform NN as a point forecaster on benchmark tests with tabular data. Recently, probabilistic forecasts in gradient boosting decision trees have been studied in (Duan et al, 2020;Sprangers et al, 2021;Brophy and Lowd, 2022). While (Duan et al, 2020;Sprangers et al, 2021) require fitting a parametric distribution (such as Gaussian or Weibull) to the data, (Brophy and Lowd, 2022) allows to produce a more flexible, nonparametric distributional forecast.…”
Section: Related Workmentioning
confidence: 99%
“…Gradient boosting decision trees are among the most popular ML models, which often outperform NN as a point forecaster on benchmark tests with tabular data. Recently, probabilistic forecasts in gradient boosting decision trees have been studied in (Duan et al, 2020;Sprangers et al, 2021;Brophy and Lowd, 2022). While (Duan et al, 2020;Sprangers et al, 2021) require fitting a parametric distribution (such as Gaussian or Weibull) to the data, (Brophy and Lowd, 2022) allows to produce a more flexible, nonparametric distributional forecast.…”
Section: Related Workmentioning
confidence: 99%
“…Examples of parametric methods include Natural Gradient Boosting for Probabilistic Prediction (NGBoost) [17], XGBoostLSS and LightGBMLSS, which model all moments of a parametric distribution: mean, location, scale, and shape (LSS) [18,19], Catboost with Uncertainty (CBU), Probabilistic Gradient Boosting Machines (PGBM) [20], and Instance-Based Uncertainty Estimation for Gradient-Boosted Regression Trees (IBUG) [21].…”
Section: Introductionmentioning
confidence: 99%