Reporting guideline for chatbot health advice studies: the Chatbot Assessment Reporting Tool (CHART) statement
- PMID: 40747825
- PMCID: PMC12314741
- DOI: 10.1093/bjs/znaf142
Reporting guideline for chatbot health advice studies: the Chatbot Assessment Reporting Tool (CHART) statement
Abstract
The Chatbot Assessment Reporting Tool (CHART) is a reporting guideline developed to provide reporting recommendations for studies evaluating the performance of generative artificial intelligence (AI)-driven chatbots when summarizing clinical evidence and providing health advice, referred to as chatbot health advice studies. CHART was developed in several phases after performing a comprehensive systematic review to identify variation in the conduct, reporting, and method in chatbot health advice studies. Findings from the review were used to develop a draft checklist that was revised through an international, multidisciplinary, modified, asynchronous Delphi consensus process of 531 stakeholders, three synchronous panel consensus meetings of 48 stakeholders, and subsequent pilot testing of the checklist. CHART includes 12 items and 39 subitems to promote transparent and comprehensive reporting of chatbot health advice studies. These include title (subitem 1a), abstract/summary (subitem 1b), background (subitems 2a,b), model identifiers (subitems 3a,b), model details (subitems 4a-c), prompt engineering (subitems 5a,b), query strategy (subitems 6a-d), performance evaluation (subitems 7a,b), sample size (subitem 8), data analysis subitem 9a), results (subitems 10a-c), discussion (subitems 11a-c), disclosures (subitem 12a), funding (subitem 12b), ethics (subitem 12c), protocol (subitem 12d), and data availability (subitem 12e). The CHART checklist and corresponding diagram of the method were designed to support key stakeholders including clinicians, researchers, editors, peer reviewers, and readers in reporting, understanding, and interpreting the findings of chatbot health advice studies.
© The Author(s) 2025. Published by Oxford University Press on behalf of BJS Foundation Ltd, by Elsevier BV, by Annals of Family Medicine, Inc., by Springer Nature, by BMJ Publishing Group Limited, and by American Medical Association.
Figures
References
-
- Huo B, Cacciamani GE, Collins GS, McKechnie T, Lee Y, Guyatt G. Reporting standards for the use of large language model-linked chatbots for health advice. Nat Med 2023;29:2988–2988 - PubMed
-
- Huo B, McKechnie T, Ortenzi M, Lee Y, Antoniou S, Mayol J et al. Dr. GPT will see you now: the ability of large language model-linked chatbots to provide colorectal cancer screening recommendations. Health Technol (Berl) 2024;14:463–469
-
- Huo B, Marfo N, Sylla P, Calabrese E, Kumar S, Slater BJ et al. Clinical artificial intelligence: teaching a large language model to generate recommendations that align with guidelines for the surgical management of GERD. Surg Endosc 2024;38:5668–5677 - PubMed
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources