An AI assistant to help review and improve causal reasoning in epidemiological documents
- PMID: 38188038
- PMCID: PMC10767365
- DOI: 10.1016/j.gloepi.2023.100130
An AI assistant to help review and improve causal reasoning in epidemiological documents
Abstract
Drawing sound causal inferences from observational data is often challenging for both authors and reviewers. This paper discusses the design and application of an Artificial Intelligence Causal Research Assistant (AIA) that seeks to help authors improve causal inferences and conclusions drawn from epidemiological data in health risk assessments. The AIA-assisted review process provides structured reviews and recommendations for improving the causal reasoning, analyses and interpretations made in scientific papers based on epidemiological data. Causal analysis methodologies range from earlier Bradford-Hill considerations to current causal directed acyclic graph (DAG) and related models. AIA seeks to make these methods more accessible and useful to researchers. AIA uses an external script (a "Causal AI Booster" (CAB) program based on classical AI concepts of slot-filling in frames organized into task hierarchies to complete goals) to guide Large Language Models (LLMs), such as OpenAI's ChatGPT or Google's LaMDA (Bard), to systematically review manuscripts and create both (a) recommendations for what to do to improve analyses and reporting; and (b) explanations and support for the recommendations. Review tables and summaries are completed systematically by the LLM in order. For example, recommendations for how to state and caveat causal conclusions in the Abstract and Discussion sections reflect previous analyses of the Study Design and Data Analysis sections. This work illustrates how current AI can contribute to reviewing and providing constructive feedback on research documents. We believe that such AI-assisted review shows promise for enhancing the quality of causal reasoning and exposition in epidemiological studies. It suggests the potential for effective human-AI collaboration in scientific authoring and review processes.
Keywords: Artificial intelligence; Causal AI boosting; Causality; Large language models (LLMs); Review methodology.
© 2023 Published by Elsevier Inc.
Conflict of interest statement
The research presented here was supported by the author's employer, Cox Associates. Cox Associates received seed funding from the American Chemistry Council in 2022 and 2023 to develop the core ideas and technology for the AI Causal Research Assistant (AIA) and its application to reviewing documents, as described in this paper. All research questions, technical approaches and innovations, example, and conclusions are solely those of the author. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Similar articles
-
Current safeguards, risk mitigation, and transparency measures of large language models against the generation of health disinformation: repeated cross sectional analysis.BMJ. 2024 Mar 20;384:e078538. doi: 10.1136/bmj-2023-078538. BMJ. 2024. PMID: 38508682 Free PMC article.
-
Academic Surgery in the Era of Large Language Models: A Review.JAMA Surg. 2024 Apr 1;159(4):445-450. doi: 10.1001/jamasurg.2023.6496. JAMA Surg. 2024. PMID: 38353991 Review.
-
ChatGPT identifies gender disparities in scientific peer review.Elife. 2023 Nov 3;12:RP90230. doi: 10.7554/eLife.90230. Elife. 2023. PMID: 37922198 Free PMC article.
-
Comparing the Efficacy of Large Language Models ChatGPT, BARD, and Bing AI in Providing Information on Rhinoplasty: An Observational Study.Aesthet Surg J Open Forum. 2023 Sep 14;5:ojad084. doi: 10.1093/asjof/ojad084. eCollection 2023. Aesthet Surg J Open Forum. 2023. PMID: 37795257 Free PMC article.
-
Large language models are changing landscape of academic publications. A positive transformation?Cas Lek Cesk. 2024;162(7-8):294-297. Cas Lek Cesk. 2024. PMID: 38981715 Review. English.
Cited by
-
An AI assistant for critically assessing and synthesizing clusters of journal articles.Glob Epidemiol. 2025 May 23;10:100207. doi: 10.1016/j.gloepi.2025.100207. eCollection 2025 Dec. Glob Epidemiol. 2025. PMID: 40548125 Free PMC article.
-
Innovations in Medicine: Exploring ChatGPT's Impact on Rare Disorder Management.Genes (Basel). 2024 Mar 28;15(4):421. doi: 10.3390/genes15040421. Genes (Basel). 2024. PMID: 38674356 Free PMC article. Review.
-
Large Language Models and the Analyses of Adherence to Reporting Guidelines in Systematic Reviews and Overviews of Reviews (PRISMA 2020 and PRIOR).J Med Syst. 2025 Jun 12;49(1):80. doi: 10.1007/s10916-025-02212-0. J Med Syst. 2025. PMID: 40504403 Free PMC article.
References
-
- Carroll R.J., Ruppert D., Stefanski L.A., Crainiceanu C.M. 2nd ed. Chapman and Hall/CRC; New York, NY: 2006. Measurement error in nonlinear models a modern perspective. - DOI
-
- Ernst G.W., Newell A. Academic Press; New York, NY: 1969. GPS: A Case Study in Generality and Problem Solving.
-
- Gerrits R.G., Jansen T., Mulyanto J., van den Berg M.J., Klazinga N.S., Kringos D.S. Occurrence and nature of questionable research practices in the reporting of messages and conclusions in international scientific Health Services Research publications: a structured assessment of publications authored by researchers in the Netherlands. BMJ Open. 2019 May 15;9(5) doi: 10.1136/bmjopen-2018-027903. PMID: 31097488; PMCID: PMC6530378. - DOI - PMC - PubMed
LinkOut - more resources
Full Text Sources