Examining Real-World Medication Consultations and Drug-Herb Interactions: ChatGPT Performance Evaluation

Background Since OpenAI released ChatGPT, with its strong capability in handling natural tasks and its user-friendly interface, it has garnered significant attention. Objective A prospective analysis is required to evaluate the accuracy and appropriateness of medication consultation responses genera...

Full description

Saved in:
Bibliographic Details
Published inJMIR medical education Vol. 9; p. e48433
Main Authors Hsu, Hsing-Yu, Hsu, Kai-Cheng, Hou, Shih-Yen, Wu, Ching-Lung, Hsieh, Yow-Wen, Cheng, Yih-Dih
Format Journal Article
LanguageEnglish
Published Toronto JMIR Publications 21.08.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Background Since OpenAI released ChatGPT, with its strong capability in handling natural tasks and its user-friendly interface, it has garnered significant attention. Objective A prospective analysis is required to evaluate the accuracy and appropriateness of medication consultation responses generated by ChatGPT. Methods A prospective cross-sectional study was conducted by the pharmacy department of a medical center in Taiwan. The test data set comprised retrospective medication consultation questions collected from February 1, 2023, to February 28, 2023, along with common questions about drug-herb interactions. Two distinct sets of questions were tested: real-world medication consultation questions and common questions about interactions between traditional Chinese and Western medicines. We used the conventional double-review mechanism. The appropriateness of each response from ChatGPT was assessed by 2 experienced pharmacists. In the event of a discrepancy between the assessments, a third pharmacist stepped in to make the final decision. Results Of 293 real-world medication consultation questions, a random selection of 80 was used to evaluate ChatGPT’s performance. ChatGPT exhibited a higher appropriateness rate in responding to public medication consultation questions compared to those asked by health care providers in a hospital setting (31/51, 61% vs 20/51, 39%; P=.01). Conclusions The findings from this study suggest that ChatGPT could potentially be used for answering basic medication consultation questions. Our analysis of the erroneous information allowed us to identify potential medical risks associated with certain questions; this problem deserves our close attention.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:2369-3762
2369-3762
DOI:10.2196/48433