2023
DOI: 10.1016/j.xops.2023.100324
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating the Performance of ChatGPT in Ophthalmology

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
29
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 189 publications
(40 citation statements)
references
References 15 publications
1
29
0
Order By: Relevance
“…Umer and Habib (2022) argue that an acceptable threshold for accuracy in diagnostic tasks should be set at >90%. However, our result is similar to other studies (Antaki et al., 2023; Gilson et al., 2023; Huh, 2023; Kung et al., 2023; Thirunavukarasu et al., 2023), that have recognized the promising potential of this LLM for medical education and clinical decision‐making. However, some studies have warned of the variability in response accuracy, with incomplete or incorrect answers being common (Lahat et al., 2023; Samaan et al., 2023).…”
Section: Discussionsupporting
confidence: 92%
See 2 more Smart Citations
“…Umer and Habib (2022) argue that an acceptable threshold for accuracy in diagnostic tasks should be set at >90%. However, our result is similar to other studies (Antaki et al., 2023; Gilson et al., 2023; Huh, 2023; Kung et al., 2023; Thirunavukarasu et al., 2023), that have recognized the promising potential of this LLM for medical education and clinical decision‐making. However, some studies have warned of the variability in response accuracy, with incomplete or incorrect answers being common (Lahat et al., 2023; Samaan et al., 2023).…”
Section: Discussionsupporting
confidence: 92%
“…In recent months, there has been a growing interest in the application of LLMs in medicine, particularly in exploring their clinical utility, as evidenced by the emergence of ChatGPT (Antaki et al, 2023;Ge & Lai, 2023;Lahat et al, 2023). Despite the promising results demonstrated by these models, it is crucial to perform a comprehensive evaluation of their performance and potential errors before determining their viability in a clinical setting (Antaki et al, 2023). In this context, a study was conducted to evaluate the consistency and accuracy of answers provided by ChatGPT to questions related to clinical situations in endodontics.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, it did not perform as well as medical students in Korea on parasitology, 9 and it achieved only 55.8% and 42.7% accuracy on high-stakes Ophthalmic Knowledge Assessment Program exams. 10 ChatGPT did not reach the passing threshold for any of the life support exams, but its answers were generally relevant and accurate and showed better congruence with resuscitation guidelines than those of similar AI systems in previous studies. 11…”
Section: Introductionmentioning
confidence: 57%
“…ChatGPT has shown various abilities in the medical field, performing better in general medicine. 10 Therefore, in this article, we tested ChatGPT with Taiwan’s Family Medicine Board Exam. The process of medical education in Taiwan involves 6 years of study in medical school, followed by the national physician licensing exam to obtain a medical license.…”
Section: Introductionmentioning
confidence: 99%