2023
DOI: 10.1101/2023.01.23.23284735
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Putting ChatGPT’s Medical Advice to the (Turing) Test

Abstract: Importance: Chatbots could play a role in answering patient questions, but patients' ability to distinguish between provider and chatbot responses, and patients' trust in chatbots' functions are not well established. Objective: To assess the feasibility of using ChatGPT or a similar AI-based chatbot for patient-provider communication. Design: Survey in January 2023 Setting: Survey Participants: A US representative sample of 400 study participants aged 18 and above was recruited on Prolific, a crowdsourcing pla… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
38
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 46 publications
(39 citation statements)
references
References 11 publications
(15 reference statements)
1
38
0
Order By: Relevance
“…However, the accuracy of computational models’ answers to medical questions is yet to exceed that of fully trained physicians, with findings in the present context of primary care being no exception [ 16 , 17 ]. When ChatGPT is used as a medical advice chatbot, advice seekers are only able to identify that the source of provided advice is computational 65% of the time [ 19 ]. It follows that health care providers must protect their patients from inaccurate information provided by this technology, as they are unable to differentiate between computational and human advice [ 19 ].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…However, the accuracy of computational models’ answers to medical questions is yet to exceed that of fully trained physicians, with findings in the present context of primary care being no exception [ 16 , 17 ]. When ChatGPT is used as a medical advice chatbot, advice seekers are only able to identify that the source of provided advice is computational 65% of the time [ 19 ]. It follows that health care providers must protect their patients from inaccurate information provided by this technology, as they are unable to differentiate between computational and human advice [ 19 ].…”
Section: Discussionmentioning
confidence: 99%
“…When ChatGPT is used as a medical advice chatbot, advice seekers are only able to identify that the source of provided advice is computational 65% of the time [ 19 ]. It follows that health care providers must protect their patients from inaccurate information provided by this technology, as they are unable to differentiate between computational and human advice [ 19 ]. This requirement for oversight limits the potential of LLMs to meaningfully change practice, as performance equivalent to that of experts is the minimum standard to justify autonomous deployment: there must be confidence in the accuracy and trustworthiness of answers from these applications [ 20 , 21 ].…”
Section: Discussionmentioning
confidence: 99%
“…We identified only one study with a similar design to ours. 18 The authors extracted patient-provider communication from electronic health records and presented five cases with provider-written answers and another five with ChatGPT-generated answers. Participants, who were recruited online, could identify 65% of AI-generated answers correctly, which is of similar magnitude to our findings, although the authors neither have formally tested any hypotheses, nor presented precision estimates for the results.…”
Section: Discussionmentioning
confidence: 99%
“…2 Some preliminary work in the medical domain highlighted ChatGPT's ability to write realistic scientific abstracts, 3 pass medical licensing exams, 4 and accurately determine appropriate radiology studies. 5 Although ChatGPT can triage medical cases, 6 answer clinical questions consistent with the judgment of practicing physicians, 7 and provide medical advice that is perceived as human-like by non-clinicians, 8 its ability to provide appropriate and equitable advice to patients across a range of clinical contexts remain unknown. These knowledge gaps are important because the underlying training data and approach for ChatGPT have not been released, 9 and there are substantive concerns about the safety, fairness, and regulation of LLMs and clinical AI systems.…”
Section: Mainmentioning
confidence: 99%