2021
DOI: 10.1177/09622802211046388
|View full text |Cite
|
Sign up to set email alerts
|

Developing clinical prediction models when adhering to minimum sample size recommendations: The importance of quantifying bootstrap variability in tuning parameters and predictive performance

Abstract: Recent minimum sample size formula (Riley et al.) for developing clinical prediction models help ensure that development datasets are of sufficient size to minimise overfitting. While these criteria are known to avoid excessive overfitting on average, the extent of variability in overfitting at recommended sample sizes is unknown. We investigated this through a simulation study and empirical example to develop logistic regression clinical prediction models using unpenalised maximum likelihood estimation, and v… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 11 publications
(14 citation statements)
references
References 40 publications
0
14
0
Order By: Relevance
“…LASSO was used because this performs variable selection (from the list of candidate predictors) through the shrinkage of the model coefficients, while also further helping to minimize overfitting. 26 We evaluated the performance of the model using calibration (agreement between the observed and expected outcomes, across the full risk range) and discrimination (ability of the model to differentiate those who had the outcome from those who did not). Calibration was assessed visually using calibration plots and summarized using the calibration-in-the-large and calibration slope.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…LASSO was used because this performs variable selection (from the list of candidate predictors) through the shrinkage of the model coefficients, while also further helping to minimize overfitting. 26 We evaluated the performance of the model using calibration (agreement between the observed and expected outcomes, across the full risk range) and discrimination (ability of the model to differentiate those who had the outcome from those who did not). Calibration was assessed visually using calibration plots and summarized using the calibration-in-the-large and calibration slope.…”
Section: Discussionmentioning
confidence: 99%
“…Model hyperparameters were determined by using grid search and selected using 10-fold cross-validation within the training data. LASSO was used because this performs variable selection (from the list of candidate predictors) through the shrinkage of the model coefficients, while also further helping to minimize overfitting 26 …”
Section: Methodsmentioning
confidence: 99%
“…A new prediction model was not constructed using the Dutch data set, as the number of cases and non‐cases was relatively low. This could lead to overfitting, rendering the improved or new model not useful for future predictions 35 …”
Section: Discussionmentioning
confidence: 99%
“…However, studies have shown that even penalization methods such as LASSO and elastic net are unstable in their selection of predictors (e.g., due to uncertainty in estimation of the tuning parameters from the data, or having a set of highly correlated predictors (Leeuwenberg et al, 2022)), especially in small sample sizes where they are arguably most needed, leading to miscalibration of predictions in new data Van Calster et al, 2020;Van Houwelingen, 2001). For this reason, Martin et al (2021) recommend model developers should use bootstrapping to investigate the uncertainty in their model's penalty terms (shrinkage factors) and predictive performance. Others studies have emphasized using methods to improve stability in the penalization approach (Roberts & Nowak, 2014), such as repeat k-fold cross-validation to estimate penalty or tuning factors from the data Seibold et al, 2018), or ensemble methods that incorporate boosting or bagging (e.g., XGBoost, random forests) (Breiman, 1996).…”
Section: Previous Researchmentioning
confidence: 99%