2023
DOI: 10.1101/2023.05.04.23289482
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

GPT-4 outperforms ChatGPT in answering non-English questions related to cirrhosis

Abstract: Background and Objectives: Artificial intelligence is increasingly being employed in healthcare, raising concerns about the exacerbation of disparities. This study evaluates ChatGPT and GPT-4's ability to comprehend and respond to cirrhosis-related questions in English, Korean, Mandarin, and Spanish, addressing language barriers that may impact patient care. Methods: A set of 36 cirrhosis-related questions were translated into Korean, Mandarin, and Spanish and prompted to both ChatGPT and GPT-4 models. Non-Eng… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
16
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 17 publications
(16 citation statements)
references
References 14 publications
0
16
0
Order By: Relevance
“…14 Another study attempted the family medicine exam using ChatGPT and was successful, partly because the questions used were from an after-school exam for medical students and likely more accessible. 7 Second, ChatGPT's language database training may not contain enough information in traditional Chinese, 15,16 In the dynamic medical field, the inability to retrieve the latest information can lead to inaccurate ChatGPT responses.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…14 Another study attempted the family medicine exam using ChatGPT and was successful, partly because the questions used were from an after-school exam for medical students and likely more accessible. 7 Second, ChatGPT's language database training may not contain enough information in traditional Chinese, 15,16 In the dynamic medical field, the inability to retrieve the latest information can lead to inaccurate ChatGPT responses.…”
Section: Discussionmentioning
confidence: 99%
“…14 Another study attempted the family medicine exam using ChatGPT and was successful, partly because the questions used were from an after-school exam for medical students and likely more accessible. 7 Second, ChatGPT’s language database training may not contain enough information in traditional Chinese, 15,16 leading to reduced accuracy and correctness in the answers when only Chinese questions are presented. Third, the Family Medicine Board Exam questions are mainly from the TAFM’s publications, including Chinese family medicine magazines and three major textbooks in the new 2023 edition ( Family Medicine , Community Medicine , and Family Doctor Clinical Practice , all in Chinese), which are not open access and not likely included in ChatGPT’s training database; this hinders the search for the most accurate answers.…”
Section: Discussionmentioning
confidence: 99%
“…The authors were primarily affiliated with institutions in the United States (n=47 of 122 different countries identified per publication, 38.5%), followed by Germany (n=11/122, 9%), Turkey (n=7/122, 5.7%), the United Kingdom (n=6/122, 4.9%), China/Australia/Italy (n=5/122, 4.1%, respectively), and 24 (n=36/122, 29.5%) other countries. Most studies examined one or more applications based on the GPT-3.5 architecture (n=66 of 124 different LLMs examined per study, 53.2%) 13,[26][27][28][29][31][32][33][34][36][37][38][39][40][42][43][44][45][46][47][48][49][52][53][54][56][57][58][59][60][61]63,[65][66][67]71,72,74,75,77,78,[81][82][83][84][85][86][87]...…”
Section: Characteristics Of Included Studiesmentioning
confidence: 99%
“…Most studies examined one or more applications based on the GPT-3.5 architecture (n=66 of 124 different LLMs examined per study, 53.2%) 13,[26][27][28][29][31][32][33][34][36][37][38][39][40][42][43][44][45][46][47][48][49][52][53][54][56][57][58][59][60][61]63,[65][66][67]71,72,74,75,77,78,[81][82][83][84][85][86][87][88][89]91,92,…”
Section: Characteristics Of Included Studiesmentioning
confidence: 99%
“…Moreover, ChatGPT-4 provides response with medical accuracy in different languages (English, Korean, Mandarin, Spanish and many others), reaching a larger population and avoiding the need for translator software that could misspell medical language and make errors while translating [34].…”
Section: Enhancing Patient-physician Interactionsmentioning
confidence: 99%