2024
DOI: 10.1002/jrsm.1710
|View full text |Cite
|
Sign up to set email alerts
|

Data extraction for evidence synthesis using a large language model: A proof‐of‐concept study

Gerald Gartlehner,
Leila Kahwati,
Rainer Hilscher
et al.

Abstract: Data extraction is a crucial, yet labor‐intensive and error‐prone part of evidence synthesis. To date, efforts to harness machine learning for enhancing efficiency of the data extraction process have fallen short of achieving sufficient accuracy and usability. With the release of large language models (LLMs), new possibilities have emerged to increase efficiency and accuracy of data extraction for evidence synthesis. The objective of this proof‐of‐concept study was to assess the performance of an LLM (Claude 2… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 25 publications
0
3
0
Order By: Relevance
“…LLMs have shown high accuracy in screening relevant titles, abstracts, and full-text 22,23 , and effectively conducting quality assessment and risk-of-bias evaluation 24 . Moreover, the LLM system has been employed to automate the extraction of Population, Intervention, Comparator, and Outcome (PICO) elements 25,26 , the generation of evidence via data extraction 27,28 , and the conduction of the network meta-analyses using generated R scripts and extracted data elements 29 . However, there has been limited exploration into the holistic integration of all steps 26 , including querying relevant articles from databases like PubMed, screening abstracts and full texts based on the PICO eligibility criteria user-defined, and extracting user-specified data elements into a computable format for downstream analysis.…”
Section: Introductionmentioning
confidence: 99%
“…LLMs have shown high accuracy in screening relevant titles, abstracts, and full-text 22,23 , and effectively conducting quality assessment and risk-of-bias evaluation 24 . Moreover, the LLM system has been employed to automate the extraction of Population, Intervention, Comparator, and Outcome (PICO) elements 25,26 , the generation of evidence via data extraction 27,28 , and the conduction of the network meta-analyses using generated R scripts and extracted data elements 29 . However, there has been limited exploration into the holistic integration of all steps 26 , including querying relevant articles from databases like PubMed, screening abstracts and full texts based on the PICO eligibility criteria user-defined, and extracting user-specified data elements into a computable format for downstream analysis.…”
Section: Introductionmentioning
confidence: 99%
“…10 The efficacy of LLMs for screening and data extraction for systematic reviews has been highly mixed, but there appear to be scenarios where they can help with these tasks. 11,12 We quantified the agreement of five individual LLMs with human consensus in the assessment of evidence appraisal tools of different levels of complexity: reporting (PRISMA) and methodological rigor (AMSTAR) of systematic reviews, and degree of pragmatism of clinical trials (PRECIS-2). We assessed how much complexity of assessment can be handled by current models, which one performs best, and whether the combination of multiple LLMs increases accuracy.…”
Section: Introductionmentioning
confidence: 99%
“…10 The efficacy of LLMs for screening and data extraction for systematic reviews has been highly mixed, but there appear to be scenarios where they can help with these tasks. 11,12…”
Section: Introductionmentioning
confidence: 99%