ChatGPT-4 accuracy for patient education in laryngopharyngeal reflux
- PMID: 38492008
- DOI: 10.1007/s00405-024-08560-w
ChatGPT-4 accuracy for patient education in laryngopharyngeal reflux
Abstract
Introduction: Chatbot Generative Pre-trained Transformer (ChatGPT) is an artificial intelligence-powered language model chatbot able to help otolaryngologists in practice and research. The ability of ChatGPT in generating patient-centered information related to laryngopharyngeal reflux disease (LPRD) was evaluated.
Methods: Twenty-five questions dedicated to definition, clinical presentation, diagnosis, and treatment of LPRD were developed from the Dubai definition and management of LPRD consensus and recent reviews. Questions about the four aforementioned categories were entered into ChatGPT-4. Four board-certified laryngologists evaluated the accuracy of ChatGPT-4 with a 5-point Likert scale. Interrater reliability was evaluated.
Results: The mean scores (SD) of ChatGPT-4 answers for definition, clinical presentation, additional examination, and treatments were 4.13 (0.52), 4.50 (0.72), 3.75 (0.61), and 4.18 (0.47), respectively. Experts reported high interrater reliability for sub-scores (ICC = 0.973). The lowest performances of ChatGPT-4 were on answers about the most prevalent LPR signs, the most reliable objective tool for the diagnosis (hypopharyngeal-esophageal multichannel intraluminal impedance-pH monitoring (HEMII-pH)), and the criteria for the diagnosis of LPR using HEMII-pH.
Conclusion: ChatGPT-4 may provide adequate information on the definition of LPR, differences compared to GERD (gastroesophageal reflux disease), and clinical presentation. Information provided upon extra-laryngeal manifestations and HEMII-pH may need further optimization. Regarding the recent trends identifying increasing patient use of internet sources for self-education, the findings of the present study may help draw attention to ChatGPT-4's accuracy on the topic of LPR.
Keywords: Artificial intelligence; ChatGPT; Chatbot; Head neck surgery; Laryngopharyngeal; Otolaryngology; Reference; Reflux.
© 2024. The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature.
References
-
- Briganti G (2023) How ChatGPT works: a mini review. Eur Arch Otorhinolaryngol. https://doi.org/10.1007/s00405-023-08337-7 - DOI - PubMed
-
- Vaira LA, Lechien JR, Abbate V, Allevi F, Audino G, Beltramini GA et al (2023) Accuracy of ChatGPT-generated information on head and neck and oromaxillofacial surgery: a multicenter collaborative analysis. Otolaryngol Head Neck Surg. https://doi.org/10.1002/ohn.489 - DOI - PubMed
-
- Davis RJ, Ayo-Ajibola O, Lin ME, Swanson MS, Chambers TN, Kwon DI, Kokot NC (2023) Evaluation of oropharyngeal cancer information from revolutionary artificial intelligence Chatbot. Laryngoscope. https://doi.org/10.1002/lary.31191 - DOI - PubMed - PMC
-
- Lechien JR, Vaezi MF, Chan WW, Allen JE, Karkos PD, Saussez S et al (2023) The Dubai definition and diagnostic criteria of laryngopharyngeal reflux: the IFOS consensus. Laryngoscope. https://doi.org/10.1002/lary.31134 - DOI - PubMed
-
- Lechien JR, Akst LM, Hamdan AL, Schindler A, Karkos PD, Barillari MR, Calvo-Henriquez C, Crevier-Buchman L, Finck C, Eun YG, Saussez S, Vaezi MF (2019) Evaluation and management of laryngopharyngeal reflux disease: state of the art review. Otolaryngol Head Neck Surg 160(5):762–782. https://doi.org/10.1177/0194599819827488 - DOI - PubMed
MeSH terms
LinkOut - more resources
Full Text Sources
