2019
DOI: 10.48550/arxiv.1910.03225
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

NGBoost: Natural Gradient Boosting for Probabilistic Prediction

Abstract: We present Natural Gradient Boosting (NG-Boost), an algorithm which brings probabilistic prediction capability to gradient boosting in a generic way. Predictive uncertainty estimation is crucial in many applications such as healthcare and weather forecasting. Probabilistic prediction, which is the approach where the model outputs a full probability distribution over the entire outcome space, is a natural way to quantify those uncertainties. Gradient Boosting Machines have been widely successful in prediction t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
17
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(17 citation statements)
references
References 14 publications
0
17
0
Order By: Relevance
“…Our aim is to learn a data-driven mapping from the set of all x i s to the set of all y i s, to be able to predict y new for a new, unseen observation x new . To solve this regression problem 5 , we employ the NGBoost algorithm (Duan et al 2019). Unlike most commonly used ML algorithms and libraries such as Random Forests (Breiman 2001), Randomized Trees (Geurts et al 2006), XGBoost (Chen & Guestrin 2016) and Gradient Boosting Machines (Ke et al 2017), NGBoost enables us to easily work in a probabilistic setting, and corresponding to every input galaxy SED, output both a measure of the central tendency (i.e.…”
Section: Proposed Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Our aim is to learn a data-driven mapping from the set of all x i s to the set of all y i s, to be able to predict y new for a new, unseen observation x new . To solve this regression problem 5 , we employ the NGBoost algorithm (Duan et al 2019). Unlike most commonly used ML algorithms and libraries such as Random Forests (Breiman 2001), Randomized Trees (Geurts et al 2006), XGBoost (Chen & Guestrin 2016) and Gradient Boosting Machines (Ke et al 2017), NGBoost enables us to easily work in a probabilistic setting, and corresponding to every input galaxy SED, output both a measure of the central tendency (i.e.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…We make the assumption that samples are drawn from Gaussian distributions. Our loss function is the negative likelihood function (Duan et al 2019).…”
Section: Proposed Methodsmentioning
confidence: 99%
“…Network), and a decision-tree based model using natural gradient boosting (NGBoost) assuming a Gaussian output distribution [24]. Hyperparameters are provided in the appendix.…”
Section: Methodsmentioning
confidence: 99%
“…Our experiments use datasets from the UCI Machine Learning Repository, and follow the same protocol as NGBoost (Hernández-Lobato and Adams, 2015;Duan et al, 2019). For all datasets, we hold out a random 10% of the examples as a test set.…”
Section: Methodsmentioning
confidence: 99%
“…These implementations can train models with hundreds of trees using millions of training examples in a matter of minutes. NGBoost (Duan et al, 2019) generalized Natural Gradient as the direction of the steepest ascent in Riemannian space, and applied it for boosting to enable the probabilistic predication capability for the regression tasks. Natural gradient boosting shows promising performance improvements on small datasets due to better training dynamics, but it suffers from slow training speed overhead especially for large datasets.…”
Section: Introductionmentioning
confidence: 99%