2020
DOI: 10.48550/arxiv.2003.07398
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Variable selection with multiply-imputed datasets: choosing between stacked and grouped methods

Abstract: Penalized regression methods, such as lasso and elastic net, are used in many biomedical applications when simultaneous regression coefficient estimation and variable selection is desired. However, missing data complicates the implementation of these methods, particularly when missingness is handled using multiple imputation. Applying a variable selection algorithm on each imputed dataset will likely lead to different sets of selected predictors, making it difficult to ascertain a final active set without reso… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…We will assess this flexible model in our sensitivity analysis. To avoid overfitting, we will employ the elastic net for a penalized estimation of regression coefficients [28,29]. The elastic net allows both selection and penalization of main effects by introducing two tuning parameters.…”
Section: Model Developmentmentioning
confidence: 99%
“…We will assess this flexible model in our sensitivity analysis. To avoid overfitting, we will employ the elastic net for a penalized estimation of regression coefficients [28,29]. The elastic net allows both selection and penalization of main effects by introducing two tuning parameters.…”
Section: Model Developmentmentioning
confidence: 99%
“…Although various models were proposed to solve the inconsistency of variable selection in MI settings, it was the first time that MI-LASSO [17] introduced the Group-LASSO [24] penalty into this problem. Different from stacking methods [10,11,15] which"stacks" multiply-imputed data as a single dataset and applies weighted models to select important variables, MI-LASSO treats the same variable across all imputed sets as a group of variables, and adopts Group-LASSO to jointly include or exclude the group of variables together. To make it clear, the mathematical loss function for Group-LASSO is:…”
Section: Mi-lassomentioning
confidence: 99%
“…Therefore information based criterion was used to select x % credible interval for shrinkage Bayesian MI-LASSO and evaluated performances. We took advantages of the modified version of Bayesian Information Criterion (BIC, referring to formula (10) and ( 11)) to assess different selection of credible interval. Also, this modified BIC was used to evaluate the five Bayesian MI-LASSO models on data.…”
Section: Performancesmentioning
confidence: 99%
See 1 more Smart Citation