2023
DOI: 10.1016/j.fertnstert.2023.05.151
|View full text |Cite
|
Sign up to set email alerts
|

The promise and peril of using a large language model to obtain clinical information: ChatGPT performs strongly as a fertility counseling tool with limitations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
23
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 34 publications
(27 citation statements)
references
References 30 publications
2
23
0
Order By: Relevance
“…It also performed well in addressing radiation oncology physics exam questions [47]. Likewise, “ChatGPT would have been at the 87 th percentile of Bunting’s 2013 international cohort for the Cardiff Fertility Knowledge Scale and at the 95 th percentile on the basis of Kudesia’s 2017 cohort for the Fertility and Infertility Treatment Knowledge Score” [48]. In addition, ChatGPT showed promising results in a simulated Ophthalmic Knowledge Assessment Program (OKAP) exam [49].…”
Section: Resultsmentioning
confidence: 99%
“…It also performed well in addressing radiation oncology physics exam questions [47]. Likewise, “ChatGPT would have been at the 87 th percentile of Bunting’s 2013 international cohort for the Cardiff Fertility Knowledge Scale and at the 95 th percentile on the basis of Kudesia’s 2017 cohort for the Fertility and Infertility Treatment Knowledge Score” [48]. In addition, ChatGPT showed promising results in a simulated Ophthalmic Knowledge Assessment Program (OKAP) exam [49].…”
Section: Resultsmentioning
confidence: 99%
“…An overview of the presence of codes for each study is provided in Supplementary Section 3. The majority of articles investigated the use and feasibility of LLMs as medical chatbots (n=84/89, 94.4%) 13,24–62,6466,68,69,7196,98111 , while fewer reports additionally or exclusively focused on the generation of patient information (n=19/89, 21.4%) 24,31,43,48,49,57,59,62,67,70,79,8891,97,102,106,107 , including clinical documentation such as informed consent forms (n=5/89, 5.6%) 43,67,91,97,102 and discharge instructions (n=1/89, 1.1%) 31 , or translation/summarization tasks of medical texts (n=5/89, 5.6%) 24,49,57,79,89 , creation of patient education materials (n=5/89, 5.6%) 48,62,90,106,107 , and simplification of radiology reports (n=2/89, 2.3%) 59,88 . Most reports evaluated LLMs in English (n=88/89, 98.9%) 13,24103,105111 , followed by Arabic (n=2/84, 2.3%) 32,104 , Mandarin (n=2/84, 2.3%) 36,75 , and Korean or Spanish (n=1/89, 1.1%, respectively) 75 .…”
Section: Resultsmentioning
confidence: 99%
“…In terms of design limitations, many authors noted the limitation that LLMs are not optimized for medical use (n=46/89, 51.7%) 13,26,28,34,35,3739,46,49,50,5459,61,62,65,66,68,70,71,7981,8385,88,91,9398,100107,109 , including implicit knowledge/lack of clinical context (n=13/89, 14.6%) 28,39,46,66,71,79,81,8385,98,103 , limitations in clinical reasoning (n=7/89, 7.9%) 55,84,95,102105 , limitations in medical image processing/production (n=5/89, 5.6%) 37,55,91,106,107 , and misunderstanding of medical information and terms by the model (n=7/89, 7.9%) 28,38,39,59,62,65,97 . In addition, data-related limitations were identified, including limited access to data on the internet (n=22/89, 24.7%) 38,39,41,43,5457,59,60,64,76,79,8284,88,91,94,96,104,109 , the undisclosed origin of training data (n=36/89, 40.5%) 25,26,29,30,32,34,36,37,40,46,…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…LLMs have also shown potential to provide responses to real-world health questions. When presented with to frequently asked clinical queries, ChatGPT is able to produce relevant, meaningful responses comparable to established sources [20,21 ▪▪ ]. One study even showed that a panel of licensed healthcare professionals preferred ChatGPT's responses to patient questions 79% of the time [22].…”
Section: Text Of Reviewmentioning
confidence: 99%