2021
DOI: 10.1101/2021.02.11.20196766
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep learning models for COVID-19 chest x-ray classification: Preventing shortcut learning using feature disentanglement

Abstract: In response to the COVID-19 global pandemic, recent research has proposed creating deep learning based models that use chest radiographs (CXRs) in a variety of clinical tasks to help manage the crisis. However, the size of existing datasets of CXRs from COVID-19+ patients are relatively small, and researchers often pool CXR data from multiple sources, for example, using different x-ray machines in various patient populations under different clinical scenarios. Deep learning models trained on such datasets have… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
20
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(20 citation statements)
references
References 20 publications
(24 reference statements)
0
20
0
Order By: Relevance
“…But the most general risk to all such compilations is confounding by dataset identity. This effect was for example clearly demonstrated for the COVIDx dataset by means of an external test set in Robinson et al. (2021) and more general for different COVID-19 related dataset merges in Ahmed, Goldgof, Paul, Goldgof, Hall, 2021 , Ahmed, Hall, Goldgof, Goldgof, Paul .…”
Section: Resultsmentioning
confidence: 74%
See 2 more Smart Citations
“…But the most general risk to all such compilations is confounding by dataset identity. This effect was for example clearly demonstrated for the COVIDx dataset by means of an external test set in Robinson et al. (2021) and more general for different COVID-19 related dataset merges in Ahmed, Goldgof, Paul, Goldgof, Hall, 2021 , Ahmed, Hall, Goldgof, Goldgof, Paul .…”
Section: Resultsmentioning
confidence: 74%
“…But many authors utilise small private dataset together with larger public datasets. An exemplary case is the use of private datasets as external test data, therewith assessing the transportability of models trained on (merged) public datasets to the private test data population and thus to the underlying hospital setting ( Park, Kim, Oh, Seo, Lee, Kim, Moon, Lim, Ye , Kim, Park, Oh, Seo, Lee, Kim, Moon, Lim, Ye , Elgendi, Nasir, Tang, Smith, Grenier, Batte, Spieler, Leslie, Menon, Fletcher, et al., 2021 , Robinson, Trivedi, Blazes, Ortiz, Desbiens, Gupta, Dodhia, Bhatraju, Liles, Lee, et al., 2021 ).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Augmentation, rather than re-training, requires less data because the augmentation model is trained on the deterioration index output (rather than its inputs), resulting in a smaller feature set during training. When using a large variety of different data sources, multimodal models trained from scratch tend to overfit to limited data [ 19 ]. By using only the outputs of existing validated models, we limit the number of features used in the image-augmented deterioration indices, mitigating the risk of overfitting.…”
Section: Discussionmentioning
confidence: 99%
“…This presents a nearly insurmountable hurdle for collaborative data sharing between institutions. Additionally, there exists a possibility that a model trained solely on medical data available in its own 'data island' is significantly overfitted, especially if all the data originates from one source [18]. This is the case for image J o u r n a l P r e -p r o o f processing algorithms, where images originating from one source may have distinct features that may lead to overfitting as training progresses.…”
Section: Introductionmentioning
confidence: 99%