2017
DOI: 10.1007/s11222-017-9754-6
|View full text |Cite
|
Sign up to set email alerts
|

Gradient boosting for distributional regression: faster tuning and improved variable selection via noncyclical updates

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
91
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6

Relationship

2
4

Authors

Journals

citations
Cited by 67 publications
(91 citation statements)
references
References 29 publications
0
91
0
Order By: Relevance
“…Additionally, this framework allowed us to incorporate smooth effects to efficiently account for spatiotemporal trends in the data that were poorly explained by other covariates and to identify those covariates and their functional forms most consistently associated with animal distribution and abundance (via stability selection). Recent advances in the application of gradient boosting (non‐cyclical application: see Thomas et al, ) could allow for even greater power in selecting appropriate variables and responses from among available covariates.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Additionally, this framework allowed us to incorporate smooth effects to efficiently account for spatiotemporal trends in the data that were poorly explained by other covariates and to identify those covariates and their functional forms most consistently associated with animal distribution and abundance (via stability selection). Recent advances in the application of gradient boosting (non‐cyclical application: see Thomas et al, ) could allow for even greater power in selecting appropriate variables and responses from among available covariates.…”
Section: Discussionmentioning
confidence: 99%
“…We fitted GAM and GAMLSS using an iterative machine‐learning approach, component‐wise functional gradient descent boosting (Bühlmann & Hothorn, ; Hothorn et al, ; Mayr et al, ; Hofner, Boccuto, & Göker, ; Mayr & Hofner, ) in a cyclical framework (Thomas et al, ). The first step of this process was to compute the negative gradient of a pre‐selected loss function, which acts as a working residual by giving more weight to observations not properly predicted in previous iterations.…”
Section: Methodsmentioning
confidence: 99%
“…Note that updating scheme (8) can also be used as a weak base learner to build a component-wise gradient boosting algorithm (Mayr et al, 2012;Thomas et al, 2018) for the fusion penalties presented in Section 3. This technique has the advantage that each model term X jk β jk can have a different amount of shrinkage and that gradient boosting does not need to compute the weights W k , since linear models are only fitted on the negative gradient −∂ℓ pen (β)/∂η θ k .…”
Section: Estimationmentioning
confidence: 99%
“…More importantly, the threshold can be chosen in such a way that the expected number of false discoveries can be theoretically bounded under mild conditions. Due to its flexibility and versatility, StabSel has received increasing popularity and been successfully applied in many domains since its inception.…”
Section: Related Workmentioning
confidence: 99%
“…A prominent feature of StabSel lies in its ability to control the FDR of the base learner through executing it multiple times on subsamples. Some scholars have argued that q Λ should be set relatively high, allowing for a certain number of false discoveries while increasing the chances that all truly important variables can be selected. Thus, we chose a false discovery tolerance of E ( V ) ≤ 4.…”
Section: Ensemble Pruning Of Stability Selectionmentioning
confidence: 99%