ChatGPT makes medicine easy to swallow: an exploratory case study on simplified radiology reports
- PMID: 37794249
- PMCID: PMC11126432
- DOI: 10.1007/s00330-023-10213-1
ChatGPT makes medicine easy to swallow: an exploratory case study on simplified radiology reports
Abstract
Objectives: To assess the quality of simplified radiology reports generated with the large language model (LLM) ChatGPT and to discuss challenges and chances of ChatGPT-like LLMs for medical text simplification.
Methods: In this exploratory case study, a radiologist created three fictitious radiology reports which we simplified by prompting ChatGPT with "Explain this medical report to a child using simple language." In a questionnaire, we tasked 15 radiologists to rate the quality of the simplified radiology reports with respect to their factual correctness, completeness, and potential harm for patients. We used Likert scale analysis and inductive free-text categorization to assess the quality of the simplified reports.
Results: Most radiologists agreed that the simplified reports were factually correct, complete, and not potentially harmful to the patient. Nevertheless, instances of incorrect statements, missed relevant medical information, and potentially harmful passages were reported.
Conclusion: While we see a need for further adaption to the medical field, the initial insights of this study indicate a tremendous potential in using LLMs like ChatGPT to improve patient-centered care in radiology and other medical domains.
Clinical relevance statement: Patients have started to use ChatGPT to simplify and explain their medical reports, which is expected to affect patient-doctor interaction. This phenomenon raises several opportunities and challenges for clinical routine.
Key points: • Patients have started to use ChatGPT to simplify their medical reports, but their quality was unknown. • In a questionnaire, most participating radiologists overall asserted good quality to radiology reports simplified with ChatGPT. However, they also highlighted a notable presence of errors, potentially leading patients to draw harmful conclusions. • Large language models such as ChatGPT have vast potential to enhance patient-centered care in radiology and other medical domains. To realize this potential while minimizing harm, they need supervision by medical experts and adaption to the medical field.
Keywords: Natural language processing; Patient-centered care; Radiology.
© 2023. The Author(s).
Conflict of interest statement
The authors of this manuscript declare no relationships with any companies whose products or services may be related to the subject matter of the article.
Figures
References
-
- ChatGPT: optimizing language models for dialogue, OpenAI (2022) [cited 2022 Dec 28]. Available via https://openai.com/blog/chatgpt/. Accessed 28 Dec 2022
-
- The Brilliance and Weirdness of ChatGPT, The New York Times (2022) Available via https://www.nytimes.com/2022/12/05/technology/chatgpt-ai-twitter.html. Accessed 28 Dec 2022
-
- Stumbling with their words, some people let AI do the talking, The Washington Post (2022) Available via https://www.washingtonpost.com/technology/2022/12/10/chatgpt-ai-helps-wr.... Accessed 28 Dec 2022
-
- ChatGPT: New AI chatbot has everyone talking to it, BBC (2022) Available via https://www.bbc.com/news/technology-63861322. Accessed 28 Dec 2022
-
- What is AI chatbot phenomenon ChatGPT and could it replace humans?, The Guardian (2022) Available via https://www.theguardian.com/technology/2022/dec/05/what-is-ai-chatbot-ph.... Accessed 28 Dec 2022
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
Other Literature Sources
