2023
DOI: 10.1093/asj/sjad128
|View full text |Cite
|
Sign up to set email alerts
|

Performance of ChatGPT on the Plastic Surgery Inservice Training Examination

Abstract: Background Developed originally as a tool for resident self-evaluation, the Plastic Surgery Inservice Training Examination (PSITE) has become a standardized tool adopted by plastic surgery residency programs. The introduction of large language models (LLMs), such as ChatGPT (OpenAI, San Francisco, CA), has demonstrated the potential to help propel the field of plastic surgery. Objectives The authors of this study wanted to as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
18
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 46 publications
(19 citation statements)
references
References 8 publications
1
18
0
Order By: Relevance
“…These results are similar to 2 studies performed in plastic surgery, reporting 57% accuracy and 55% accuracy on the 2022 Plastic Surgery In‐Service exams. 14 , 15 Our study differed from those studies in that we stratified questions based on difficulty, showing that ChatGPT may be able to answer easy questions with better proficiency, but more in‐depth, nuanced otolaryngology topics are difficult for the chatbot to correctly answer at this time. Within the field of medicine, it is of utmost importance that the tools we use as educational resources, and those that support clinical decision making, are validated.…”
Section: Discussionmentioning
confidence: 59%
“…These results are similar to 2 studies performed in plastic surgery, reporting 57% accuracy and 55% accuracy on the 2022 Plastic Surgery In‐Service exams. 14 , 15 Our study differed from those studies in that we stratified questions based on difficulty, showing that ChatGPT may be able to answer easy questions with better proficiency, but more in‐depth, nuanced otolaryngology topics are difficult for the chatbot to correctly answer at this time. Within the field of medicine, it is of utmost importance that the tools we use as educational resources, and those that support clinical decision making, are validated.…”
Section: Discussionmentioning
confidence: 59%
“…Several studies focused on the role of generative AI models in tests of medical knowledge [ 8 - 11 , 13 , 26 , 27 , 31 - 39 ]. These examinations ranged from general medical knowledge tests such as the United States Medical Licensing Exam to specialized examinations in fields like cardiology, neurology, and ophthalmology [ 8 , 9 , 33 , 37 , 38 ].…”
Section: Resultsmentioning
confidence: 99%
“…Despite not being trained on a specific data set, ChatGPT performed at the level of a first-year resident in plastic surgery on the in-service training exam. 7,8 In neurosurgery, ChatGPT performed worse than the average user on Self-Assessment Neurosurgery questions but better than residents in some topics. 9 Clearly, there is already some rudimentary capacity in providing specialty care.…”
Section: Discussionmentioning
confidence: 99%