Performance and risks of ChatGPT used in drug information: an exploratory real-world analysis
- PMID: 37263772
- DOI: 10.1136/ejhpharm-2023-003750
Performance and risks of ChatGPT used in drug information: an exploratory real-world analysis
Abstract
Objectives: To investigate the performance and risk associated with the usage of Chat Generative Pre-trained Transformer (ChatGPT) to answer drug-related questions.
Methods: A sample of 50 drug-related questions were consecutively collected and entered in the artificial intelligence software application ChatGPT. Answers were documented and rated in a standardised consensus process by six senior hospital pharmacists in the domains content (correct, incomplete, false), patient management (possible, insufficient, not possible) and risk (no risk, low risk, high risk). As reference, answers were researched in adherence to the German guideline of drug information and stratified in four categories according to the sources used. In addition, the reproducibility of ChatGPT's answers was analysed by entering three questions at different timepoints repeatedly (day 1, day 2, week 2, week 3).
Results: Overall, only 13 of 50 answers provided correct content and had enough information to initiate management with no risk of patient harm. The majority of answers were either false (38%, n=19) or had partly correct content (36%, n=18) and no references were provided. A high risk of patient harm was likely in 26% (n=13) of the cases and risk was judged low for 28% (n=14) of the cases. In all high-risk cases, actions could have been initiated based on the provided information. The answers of ChatGPT varied over time when entered repeatedly and only three out of 12 answers were identical, showing no reproducibility to low reproducibility.
Conclusion: In a real-world sample of 50 drug-related questions, ChatGPT answered the majority of questions wrong or partly wrong. The use of artificial intelligence applications in drug information is not possible as long as barriers like wrong content, missing references and reproducibility remain.
Keywords: EVIDENCE-BASED MEDICINE; HEALTH SERVICES ADMINISTRATION; JOURNALISM, MEDICAL; Medical Informatics; PHARMACY SERVICE, HOSPITAL.
© European Association of Hospital Pharmacists 2024. No commercial re-use. See rights and permissions. Published by BMJ.
Conflict of interest statement
Competing interests: None declared.
Similar articles
-
Assessing the applicability and appropriateness of ChatGPT in answering clinical pharmacy questions.Ann Pharm Fr. 2024 May;82(3):507-513. doi: 10.1016/j.pharma.2023.11.001. Epub 2023 Nov 20. Ann Pharm Fr. 2024. PMID: 37992892
-
Evaluating the performance of ChatGPT in clinical pharmacy: A comparative study of ChatGPT and clinical pharmacists.Br J Clin Pharmacol. 2024 Jan;90(1):232-238. doi: 10.1111/bcp.15896. Epub 2023 Sep 13. Br J Clin Pharmacol. 2024. PMID: 37626010
-
A comparison of drug information question responses by a drug information center and by ChatGPT.Am J Health Syst Pharm. 2025 Apr 17;82(8):448-460. doi: 10.1093/ajhp/zxae316. Am J Health Syst Pharm. 2025. PMID: 39450858
-
ChatGPT's performance in German OB/GYN exams - paving the way for AI-enhanced medical education and clinical practice.Front Med (Lausanne). 2023 Dec 13;10:1296615. doi: 10.3389/fmed.2023.1296615. eCollection 2023. Front Med (Lausanne). 2023. PMID: 38155661 Free PMC article.
-
Assessing the Performance of Chat Generative Pretrained Transformer (ChatGPT) in Answering Andrology-Related Questions.Urol Res Pract. 2023 Nov;49(6):365-369. doi: 10.5152/tud.2023.23171. Urol Res Pract. 2023. PMID: 37933835 Free PMC article.
Cited by
-
HIV Prevention and Treatment Information from Four Artificial Intelligence Platforms: A Thematic Analysis.AIDS Behav. 2025 Jun 7. doi: 10.1007/s10461-025-04786-9. Online ahead of print. AIDS Behav. 2025. PMID: 40481266
-
ChatGPT fails challenging the recent ESCMID brain abscess guideline.J Neurol. 2024 Apr;271(4):2086-2101. doi: 10.1007/s00415-023-12168-1. Epub 2024 Jan 27. J Neurol. 2024. PMID: 38279999 Free PMC article.
-
Poor performance of ChatGPT in clinical rule-guided dose interventions in hospitalized patients with renal dysfunction.Eur J Clin Pharmacol. 2024 Aug;80(8):1133-1140. doi: 10.1007/s00228-024-03687-5. Epub 2024 Apr 9. Eur J Clin Pharmacol. 2024. PMID: 38592470
-
Can We Ask ChatGPT About Drug Safety? Appropriateness of ChatGPT Responses to Questions About Drug Use and Adverse Reactions Received by Pharmacovigilance Centers.Drug Saf. 2024 Sep;47(9):921-923. doi: 10.1007/s40264-024-01437-1. Epub 2024 May 8. Drug Saf. 2024. PMID: 38717670 No abstract available.
-
Need an AI-Enabled, Next-Generation, Advanced ChatGPT or Large Language Models (LLMs) for Error-Free and Accurate Medical Information.Ann Biomed Eng. 2024 Feb;52(2):134-135. doi: 10.1007/s10439-023-03297-9. Epub 2023 Jun 27. Ann Biomed Eng. 2024. PMID: 37368124
MeSH terms
LinkOut - more resources
Full Text Sources
Research Materials