2021
DOI: 10.1371/journal.pone.0256152
|View full text |Cite
|
Sign up to set email alerts
|

Radiomics machine learning study with a small sample size: Single random training-test set split may lead to unreliable results

Abstract: This study aims to determine how randomly splitting a dataset into training and test sets affects the estimated performance of a machine learning model and its gap from the test performance under different conditions, using real-world brain tumor radiomics data. We conducted two classification tasks of different difficulty levels with magnetic resonance imaging (MRI) radiomics features: (1) “Simple” task, glioblastomas [n = 109] vs. brain metastasis [n = 58] and (2) “difficult” task, low- [n = 163] vs. high-gr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(29 citation statements)
references
References 26 publications
0
24
0
Order By: Relevance
“…The authors have not utilized a validation set and as such the proposed ML models are likely overfitted to the test set and report a misleading higher area under the curve performance. 5…”
Section: Sirmentioning
confidence: 99%
“…The authors have not utilized a validation set and as such the proposed ML models are likely overfitted to the test set and report a misleading higher area under the curve performance. 5…”
Section: Sirmentioning
confidence: 99%
“…Recent work has shown that using ML methods after a single random training-validation dataset split may yield unreliable results. 17 Thus, to assess the method's out-ofsample performance and to minimize the risk of overfitting, cross-validation offers a simple way to test the accuracy of the method for new (independent) data. For instance, k-fold cross validation procedures can provide unbiased estimates of the true generalization performance for feature selection, model comparison, or classification accuracy.…”
Section: Consider the Use Of Independent Datasets Or Cross-validationmentioning
confidence: 99%
“…Recent work has shown that using ML methods after a single random training‐validation dataset split may yield unreliable results 17 . Thus, to assess the method's out‐of‐sample performance and to minimize the risk of over‐fitting, cross‐validation offers a simple way to test the accuracy of the method for new (independent) data.…”
Section: Consider the Use Of Independent Datasets Or Cross‐validationmentioning
confidence: 99%
“…[7] In addition, a study demonstrated that even with cross-validation, studies with a small sample size are still unreliable. [38] Therefore, multi-center research is necessary to obtain sufficient sample size and avoid unstable and suboptimal results.…”
Section: Challenges and Potential Solutionsmentioning
confidence: 99%