2020
DOI: 10.1016/j.biopsych.2019.12.001
|View full text |Cite
|
Sign up to set email alerts
|

Multiple Holdouts With Stability: Improving the Generalizability of Machine Learning Analyses of Brain–Behavior Relationships

Abstract: BACKGROUND: In 2009, the National Institute of Mental Health launched the Research Domain Criteria, an attempt to move beyond diagnostic categories and ground psychiatry within neurobiological constructs that combine different levels of measures (e.g., brain imaging and behavior). Statistical methods that can integrate such multimodal data, however, are often vulnerable to overfitting, poor generalization, and difficulties in interpreting the results. METHODS: We propose an innovative machine learning framewor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
58
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8

Relationship

1
7

Authors

Journals

citations
Cited by 39 publications
(58 citation statements)
references
References 84 publications
0
58
0
Order By: Relevance
“…method has been employed successfully in prior studies to provide insight into how panels of cognitive and behavioural measures relate to multivariate neuroimaging-derived phenotypes [34][35][36] . PLS-like analyses can be problematic if not properly validated (for example producing spurious results due to overfitting), and so we adopted best-practice methods for validating these results [37][38][39] , selecting the optimal number of components using cross-validation and training the model on 75% of the data, before testing its performance on the remaining 25% of the data.…”
Section: Resultsmentioning
confidence: 99%
“…method has been employed successfully in prior studies to provide insight into how panels of cognitive and behavioural measures relate to multivariate neuroimaging-derived phenotypes [34][35][36] . PLS-like analyses can be problematic if not properly validated (for example producing spurious results due to overfitting), and so we adopted best-practice methods for validating these results [37][38][39] , selecting the optimal number of components using cross-validation and training the model on 75% of the data, before testing its performance on the remaining 25% of the data.…”
Section: Resultsmentioning
confidence: 99%
“…Generally speaking, the sample size included in D 1 should account for at least 2/3 of the D of the entire dataset. In practice, there is a widely used Hold-Out method [ 22 ]: when the data has obvious time series factors, the time of online data is after the offline dataset. In this case, the training set and test set should be divided according to time.…”
Section: Resultsmentioning
confidence: 99%
“…For each permuted sample, we performed s-PLS-DA using the optimal parameters identified in the original model and retested the weights identified in the permuted data on the held-out set. To test generalizability, we followed an established multi-holdout procedure 33 , whereby we created ten randomly selected optimization/held-out samples using the 80/20 split and repeated the above analyses. Of the ten samples, the one yielding the lowest P -value in the held-out set was chosen as the best model based on a Bonferroni corrected P < 0.05.…”
Section: Methodsmentioning
confidence: 99%