2021
DOI: 10.1177/09622802211051088
|View full text |Cite
|
Sign up to set email alerts
|

Deselection of base-learners for statistical boosting—with an application to distributional regression

Abstract: We present a new procedure for enhanced variable selection for component-wise gradient boosting. Statistical boosting is a computational approach that emerged from machine learning, which allows to fit regression models in the presence of high-dimensional data. Furthermore, the algorithm can lead to data-driven variable selection. In practice, however, the final models typically tend to include too many variables in some situations. This occurs particularly for low-dimensional data ([Formula: see text]), where… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

1
18
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

5
3

Authors

Journals

citations
Cited by 13 publications
(19 citation statements)
references
References 44 publications
1
18
0
Order By: Relevance
“…In contrast to classical regression methods, boosting does not provide closed formulas for standard errors of effect estimates or confidence intervals April 29, 2022 14/23 that could be used for inference. Furthermore, as mentioned before, statistical boosting is in general associated with a slightly higher computational complexity compared to methods such as the lasso [26] and has a known tendency to include too many variables in low-dimensional settings [19,45]. Our results suggest that the incorporation of batches substantially reduced the computational time.…”
Section: Discussionsupporting
confidence: 57%
“…In contrast to classical regression methods, boosting does not provide closed formulas for standard errors of effect estimates or confidence intervals April 29, 2022 14/23 that could be used for inference. Furthermore, as mentioned before, statistical boosting is in general associated with a slightly higher computational complexity compared to methods such as the lasso [26] and has a known tendency to include too many variables in low-dimensional settings [19,45]. Our results suggest that the incorporation of batches substantially reduced the computational time.…”
Section: Discussionsupporting
confidence: 57%
“…In contrast to classical regression methods, boosting does not provide closed formulas for standard errors of effect estimates or confidence intervals that could be used for inference. Furthermore, as mentioned before, statistical boosting is in general associated with a slightly higher computational complexity compared to methods such as the lasso ( Hepp et al, 2016 ) and has a known tendency to include too many variables in low-dimensional settings ( Staerk and Mayr, 2021 ; Strömer et al, 2022 ). Our results suggest that the incorporation of batches substantially reduced the computational time.…”
Section: Discussionmentioning
confidence: 99%
“…A limitation of our algorithm is the relatively high selection rates for variables with only minor importance, which occurs particularly in low‐dimensional settings. In this context, Strömer et al 66 have recently proposed an approach to deselect predictors with negligible impact to obtain sparser models with statistical boosting. We want to investigate the incorporation of this proposal in the context of multivariate GAMLSS in the future.…”
Section: Discussionmentioning
confidence: 99%