Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Mar 3;9(2):pkaf021.
doi: 10.1093/jncics/pkaf021.

The use of large language models to enhance cancer clinical trial educational materials

Affiliations

The use of large language models to enhance cancer clinical trial educational materials

Mingye Gao et al. JNCI Cancer Spectr. .

Abstract

Background: Adequate patient awareness and understanding of cancer clinical trials is essential for trial recruitment, informed decision making, and protocol adherence. Although large language models (LLMs) have shown promise for patient education, their role in enhancing patient awareness of clinical trials remains unexplored. This study explored the performance and risks of LLMs in generating trial-specific educational content for potential participants.

Methods: Generative Pretrained Transformer 4 (GPT4) was prompted to generate short clinical trial summaries and multiple-choice question-answer pairs from informed consent forms from ClinicalTrials.gov. Zero-shot learning was used for summaries, using a direct summarization, sequential extraction, and summarization approach. One-shot learning was used for question-answer pairs development. We evaluated performance through patient surveys of summary effectiveness and crowdsourced annotation of question-answer pair accuracy, using held-out cancer trial informed consent forms not used in prompt development.

Results: For summaries, both prompting approaches achieved comparable results for readability and core content. Patients found summaries to be understandable and to improve clinical trial comprehension and interest in learning more about trials. The generated multiple-choice questions achieved high accuracy and agreement with crowdsourced annotators. For both summaries and multiple-choice questions, GPT4 was most likely to include inaccurate information when prompted to provide information that was not adequately described in the informed consent forms.

Conclusions: LLMs such as GPT4 show promise in generating patient-friendly educational content for clinical trials with minimal trial-specific engineering. The findings serve as a proof of concept for the role of LLMs in improving patient education and engagement in clinical trials, as well as the need for ongoing human oversight.

PubMed Disclaimer

Conflict of interest statement

D.K.: Advisory and consulting, unrelated to this work: Genentech/Roche.

D.S.B.: Editorial, unrelated to this work: associate editor of Radiation Oncology, HemOnc.org (no financial compensation); advisory and consulting, unrelated to this work: MercurialAI (Dr. Danielle Bitterman has a financial interest in Mercurial AI, which is developing an AI platform to assist oncologists with reviewing patient records and treatment recommendations, and an AI chatbot to provide guidance and reassurance to cancer patients. Dr. Bitterman’s interests were reviewed and are managed by Brigham and Women’s Hospital and Mass General Brigham in accordance with their conflict of interest policies.).

H.I.: Principal, unrelated to this work; Digidence.

L.S.L.: Employee of Verily; advisor unrelated to this work: MediSensor Technologies, BellSant.

X.C.: Employee of Centaur Labs.

E.D.: Employee of Centaur Labs.

Figures

Figure 1.
Figure 1.
Illustration of the overall study design, including the approach to generating and evaluating summaries and multiple-choice question-answer pairs from clinical trial informed consent forms. Abbreviation: ICF = informed consent form; LLM = large language model; MCQA = multiple-choice question-answer.
Figure 2.
Figure 2.
Clinician evaluation of 11 clinical trial summaries generated using the sequential prompting approach (left) and the direct prompting approach (right). There were 44 responses per item in the clinician evaluation summary.
Figure 3.
Figure 3.
Clinician evaluation of overall quality of the 11 clinical trial summaries generated using the sequential prompting approach (left) and the direct prompting approach (right).

Comment in

References

    1. Kumar G, Chaudhary P, Quinn A, Su D. Barriers for cancer clinical trial enrollment: A qualitative study of the perspectives of healthcare providers. Contemp Clin Trials Commun. 2022;28:100939. 10.1016/j.conctc.2022.100939 - DOI - PMC - PubMed
    1. BECOME Initiative Final Report. MBCA (Melanoma and Brain Cancer Alliance). Retrieved from https://www.mbcalliance.org/wp-content/uploads/BECOME-Final-Report-FULL.pdf
    1. Pretesting NIH Clinical Trial Awareness Messages: A Focus Study with Patients, Caregivers, and the General Public. National Institutes of Health; 2011.
    1. The need for awareness of clinical research. National Institutes of Health (NIH). Accessed May 30, 2015. https://www.nih.gov/health-information/nih-clinical-research-trials-you/...
    1. N. L. of Medicine. ClinicalTrials.gov. 2023. Accessed November 23, 2023. https://clinicaltrials.gov/

LinkOut - more resources