2020
DOI: 10.1136/bmj.m689
|View full text |Cite
|
Sign up to set email alerts
|

Artificial intelligence versus clinicians: systematic review of design, reporting standards, and claims of deep learning studies

Abstract: Objective To systematically examine the design, reporting standards, risk of bias, and claims of studies comparing the performance of diagnostic deep learning algorithms for medical imaging with that of expert clinicians. Design Systematic review. Data sources Medline, Embase, Cochrane Central Register of Controlled Trials, and the World Health Organization trial registry from 2010 to June 2019. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

11
500
0
13

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 641 publications
(576 citation statements)
references
References 35 publications
11
500
0
13
Order By: Relevance
“…It has been recognised that most recent AI studies are inadequately reported and existing reporting guidelines do not fully cover potential sources of bias specific to AI systems 25. The welcome emergence of randomised controlled trials (RCTs) seeking to evaluate newer interventions based on, or including, an AI component (hereafter “AI interventions”)23262728293031 has similarly been met with concerns about the design and reporting 25323334. This has highlighted the need to provide reporting guidance that is “fit-for-purpose” in this domain.…”
Section: Introductionmentioning
confidence: 99%
“…It has been recognised that most recent AI studies are inadequately reported and existing reporting guidelines do not fully cover potential sources of bias specific to AI systems 25. The welcome emergence of randomised controlled trials (RCTs) seeking to evaluate newer interventions based on, or including, an AI component (hereafter “AI interventions”)23262728293031 has similarly been met with concerns about the design and reporting 25323334. This has highlighted the need to provide reporting guidance that is “fit-for-purpose” in this domain.…”
Section: Introductionmentioning
confidence: 99%
“…We will develop a data extraction instrument for study data based on several previous systematic reviews of prediction model [51][52][53]. As the reviewers have different levels of experience and knowledge, the items listed will be reviewed and discussed to ensure that all reviewers had clear knowledge of the procedures.…”
Section: Data Collection -Data Extractionmentioning
confidence: 99%
“…However, they include many items that may not available for prediction models built with traditional statistical methods based on clinical characteristics, laboratory examinations, or genetic factors. On the other hand, TROPOD, CHARMS and PROBAST have been already proved suitable for assessing prediction models using arti cial intelligence methods [53]. Thus, we will choose three more widely-adapted and more extensively-accepted tools, to develop our critical appraisal instrument.…”
Section: -Critical Appraisalmentioning
confidence: 99%
“…It has been recognised that most recent AI studies are inadequately reported and existing reporting guidelines do not fully cover potential sources of bias specific to AI systems 17. The welcome emergence of randomised controlled trials (RCTs) seeking to evaluate clinical efficacy of newer interventions based on, or including, an AI component (hereafter ‘AI interventions’) 15181920212223 has similarly been met with concerns about design and reporting 17242526. This has highlighted the need to provide reporting guidance that is ‘fit-for-purpose’ in this domain.…”
Section: Introductionmentioning
confidence: 99%