2023
DOI: 10.1590/1806-9282.20230848
|View full text |Cite
|
Sign up to set email alerts
|

Performance of ChatGPT-4 in answering questions from the Brazilian National Examination for Medical Degree Revalidation

Mauro Gobira,
Luis Filipe Nakayama,
Rodrigo Moreira
et al.

Abstract: SUMMARY OBJECTIVE: The aim of this study was to evaluate the performance of ChatGPT-4.0 in answering the 2022 Brazilian National Examination for Medical Degree Revalidation (Revalida) and as a tool to provide feedback on the quality of the examination. METHODS: A total of two independent physicians entered all examination questions into ChatGPT-4.0. After comparing the outputs with the test solutions, they classified the large language model answers as adeq… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 27 publications
(13 citation statements)
references
References 9 publications
0
10
0
Order By: Relevance
“…Results were similar among the American Academy of Ophthalmology's Basic and Clinical Science Course (46.0%-84.3%), 41,42,46,55 Ophthoquestions (42.7%-84%), 41,47,48 Fellow of The Royal College of Ophthalmologists (FRCOphth) examination questions (32%-88.4%) 51,57 and Statpearls (55.5%-73.2%). 49 In comparison, a lower score was observed in Brazil board examination questions (41.5%) 44 and higher in European board examinations (91%). 50 The performance of LLMs was found to be better for the subspecialties of medicine, cornea, refractive surgery and oncology, and weakest for glaucoma, neuro-ophthalmology, pathology, tumours, optics, oculoplastic and mathematical concepts.…”
Section: Diagnosis Informaɵonmentioning
confidence: 81%
See 2 more Smart Citations
“…Results were similar among the American Academy of Ophthalmology's Basic and Clinical Science Course (46.0%-84.3%), 41,42,46,55 Ophthoquestions (42.7%-84%), 41,47,48 Fellow of The Royal College of Ophthalmologists (FRCOphth) examination questions (32%-88.4%) 51,57 and Statpearls (55.5%-73.2%). 49 In comparison, a lower score was observed in Brazil board examination questions (41.5%) 44 and higher in European board examinations (91%). 50 The performance of LLMs was found to be better for the subspecialties of medicine, cornea, refractive surgery and oncology, and weakest for glaucoma, neuro-ophthalmology, pathology, tumours, optics, oculoplastic and mathematical concepts.…”
Section: Diagnosis Informaɵonmentioning
confidence: 81%
“…LLMs regularly outperformed the threshold of ophthalmological specialist examinations. [40][41][42][43][44][45][46][47][48][49][50][51][52][53][54][55][56][57] However, the accuracy of LLMs was 66.9% (22.4%-91%) while the ophthalmology trainees scored 68.4% (33%-75.7%-Table 3).…”
Section: Performance In Qualifying Ophthalmological Board Examination...mentioning
confidence: 99%
See 1 more Smart Citation
“…Examination in Japan [56,57], and the Brazilian National Examination for Medical Degree Revalidation [58].…”
Section: Discussionmentioning
confidence: 99%
“…On the other hand, the performance of the AI models in this study was not entirely an unexpected finding. This comes in light of the recent evidence showing AI models’ abilities to pass reputable exams in multiple languages such as the USMLE [37], the German State Examination in Medicine [55], the National Medical Licensing Examination in Japan [56, 57], and the Brazilian National Examination for Medical Degree Revalidation [58].…”
Section: Discussionmentioning
confidence: 99%