2022
DOI: 10.1371/journal.pdig.0000062
|View full text |Cite
|
Sign up to set email alerts
|

A novel interpretable machine learning system to generate clinical risk scores: An application for predicting early mortality or unplanned readmission in a retrospective cohort study

Abstract: Risk scores are widely used for clinical decision making and commonly generated from logistic regression models. Machine-learning-based methods may work well for identifying important predictors to create parsimonious scores, but such ‘black box’ variable selection limits interpretability, and variable importance evaluated from a single model can be biased. We propose a robust and interpretable variable selection approach using the recently developed Shapley variable importance cloud (ShapleyVIC) that accounts… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8
1

Relationship

2
7

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 25 publications
(56 reference statements)
0
7
0
Order By: Relevance
“…In the first step, the overall importance of all 41 variables to logistic regression analysis of PD‐MCI was assessed using ShapleyVIC to screen out variables less useful to the analysis. ShapleyVIC is a recently developed interpretable machine learning method for comprehensive variable importance analysis [19], and its clinical application has been illustrated in a study of adverse outcomes after hospital discharge [25]. Compared to conventional inference, it extends the analysis to a group of nearly optimal models and pools information across these good models using a meta‐analysis approach to provide a robust measure of overall variable importance.…”
Section: Methodsmentioning
confidence: 99%
“…In the first step, the overall importance of all 41 variables to logistic regression analysis of PD‐MCI was assessed using ShapleyVIC to screen out variables less useful to the analysis. ShapleyVIC is a recently developed interpretable machine learning method for comprehensive variable importance analysis [19], and its clinical application has been illustrated in a study of adverse outcomes after hospital discharge [25]. Compared to conventional inference, it extends the analysis to a group of nearly optimal models and pools information across these good models using a meta‐analysis approach to provide a robust measure of overall variable importance.…”
Section: Methodsmentioning
confidence: 99%
“…Logistic regression was selected because of its interpretability in understanding which features are the most important for the prediction task [5,[30][31][32][33][34], compared to more complex models such as deep neural networks [34,35]. This is a crucial characteristic of the logistic regression model for this research, given that the main objective of this paper is to determine which amino acids are most significant for MDD prediction.…”
Section: Classification Algorithmmentioning
confidence: 99%
“… 3 The modularized structure allows AutoScore to be integrated with more advanced interpretable machine learning methods (e.g., the Shapley variable importance cloud 28 ) for improved robustness, interpretability and transparency in the risk score development. 29 …”
Section: Before You Beginmentioning
confidence: 99%