Medical Imaging 2020: Image Perception, Observer Performance, and Technology Assessment 2020
DOI: 10.1117/12.2550066
|View full text |Cite
|
Sign up to set email alerts
|

Network output visualization to uncover limitations of deep learning detection of pneumothorax

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

3
5
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 0 publications
3
5
0
Order By: Relevance
“…In addition to being used for clinical interpretation, saliency method heat maps are also used for the evaluation of CXR interpretation models, for quality improvement (QI) and quality assurance (QA) in clinical practices, and for dataset annotation 51 . However, we found that saliency method localization performance, on balance, performed worse than expert localization across multiple analyses and across many important pathologies (our findings are consistent with recent work focused on localizing a single pathology, Pneumothorax, in CXRs 52 ). If used in clinical practice, heat maps that incorrectly highlight medical images may exacerbate well documented biases (chiefly, automation bias) and erode trust in model predictions (even when model output is correct), limiting clinical translation 22 .…”
Section: Discussionsupporting
confidence: 88%
See 1 more Smart Citation
“…In addition to being used for clinical interpretation, saliency method heat maps are also used for the evaluation of CXR interpretation models, for quality improvement (QI) and quality assurance (QA) in clinical practices, and for dataset annotation 51 . However, we found that saliency method localization performance, on balance, performed worse than expert localization across multiple analyses and across many important pathologies (our findings are consistent with recent work focused on localizing a single pathology, Pneumothorax, in CXRs 52 ). If used in clinical practice, heat maps that incorrectly highlight medical images may exacerbate well documented biases (chiefly, automation bias) and erode trust in model predictions (even when model output is correct), limiting clinical translation 22 .…”
Section: Discussionsupporting
confidence: 88%
“…Our work has several potential implications for patient care. Heat maps generated using saliency methods are advocated as clinical decision support in the hope that the heat maps not only improve clinical decision-making, but also encourage clinicians to trust model predictions [32][33][34] . However, we found that AI localization performance, on balance, how we might improve saliency methods in the future.…”
Section: Discussionmentioning
confidence: 99%
“…In general, the particular localization tasks presented in this paper can be incredibly difficult due to the overlapping structures present in 2D chest radiographs, as well as subtle changes in texture that can be challenging to detect [17]. The challenges of the pneumothorax and pneumonia chest radiograph datasets serve to demonstrate the limitations on localization abilities of saliency maps.…”
Section: Supplementary Materialsmentioning
confidence: 99%
“…The study found that models of similar accuracy produced different explanations, and GradCAM would even obscure most of the lesion of interest causing any explanation for melanoma classification to be clinically useless. Additionally, only two studies in the medical domain assessed saliency maps’ localization capabilities using some ground-truth measure, such as bounding boxes or semantic segmentation [17,45]. However, in Crosby et al 2020 there was no quantification of the extent of overlap (utility) of GradCAM with the relevant image regions, but rather a binary measure of whether or not GradCAM’s region of highest activation intersected the pneumothorax region.…”
mentioning
confidence: 99%
See 1 more Smart Citation