2019
DOI: 10.2139/ssrn.3384923
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning Under Scrutiny: Performance Against Health Care Professionals in Detecting Diseases from Medical Imaging - Systematic Review and Meta-Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 0 publications
1
7
0
Order By: Relevance
“…With nearly all studies providing detail on data sources, eligibility criteria and diagnosis classification, only a few studies reported study participant flow diagram, distribution of disease severity and distribution of alternative diagnosis. Our findings are in line with those of Faes et al recent systematic reviews 22 in which they found poor reporting and potential biases arising from study design in studies using ML Open access methods for classifying diseases from medical imaging. Similarly, in another systematic review, Christodoulou et al 23 found studies comparing the performance of logistic regression models with ML models for clinical prediction to have poor methodology and reporting quality.…”
Section: Discussionsupporting
confidence: 92%
“…With nearly all studies providing detail on data sources, eligibility criteria and diagnosis classification, only a few studies reported study participant flow diagram, distribution of disease severity and distribution of alternative diagnosis. Our findings are in line with those of Faes et al recent systematic reviews 22 in which they found poor reporting and potential biases arising from study design in studies using ML Open access methods for classifying diseases from medical imaging. Similarly, in another systematic review, Christodoulou et al 23 found studies comparing the performance of logistic regression models with ML models for clinical prediction to have poor methodology and reporting quality.…”
Section: Discussionsupporting
confidence: 92%
“…To repeat, we should take preliminary reports on the reliability of AI systems in medicine with a grain of salt. As of yet, it is hard to reproduce the results of various machine learning studies, which makes it difficult to convince the medical community of their reliability and accuracy (Faes et al 2019;Olorisade et al 2017). As of yet, there are no standardized performance metrics for machine learning systems, which makes it hard to compare these systems with each other and with human practitioners ( Japkowicz & Shah 2011).…”
Section: Black-box Medicinementioning
confidence: 99%
“…If a general practitioner is unsure about how to interpret, say, a MR image, it seems perfectly reasonable to say that he should adopt the opinion that an expert in MRI interpretation would form after inspecting the image. In see Faes et al (2019). As a reviewer for this journal interestingly noted, this observation might point towards a correlation between the amount of information that a clinician has and the degree to which he or she is epistemically obliged to rely on AI systems in decision making.…”
Section: Black-box Medicinementioning
confidence: 99%
“…In some medical domains, the use of AI systems will replace a considerable part of the work of human experts [12]. However, performance comparisons between AI systems and human experts suffer from the difficulty in reproducing and comparing results due to the lack of a unified approach [11,13,14]. Yet, it is reasonable to assume that AI systems will increasingly gain in epistemic authority, even though the assistive role of AI is frequently emphasized [15].…”
Section: Ai-based Decision Support In Healthcarementioning
confidence: 99%