Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
[Preprint]. 2024 May 22:rs.3.rs-4240043.
doi: 10.21203/rs.3.rs-4240043/v1.

Me-LLaMA: Foundation Large Language Models for Medical Applications

Affiliations

Me-LLaMA: Foundation Large Language Models for Medical Applications

Qianqian Xie et al. Res Sq. .

Abstract

Recent advancements in large language models (LLMs) such as ChatGPT and LLaMA have hinted at their potential to revolutionize medical applications, yet their application in clinical settings often reveals limitations due to a lack of specialized training on medical-specific data. In response to this challenge, this study introduces Me-LLaMA, a novel medical LLM family that includes foundation models - Me-LLaMA 13/70B, along with their chat-enhanced versions - Me-LLaMA 13/70B-chat, developed through continual pre-training and instruction tuning of LLaMA2 using large medical datasets. Our methodology leverages a comprehensive domain-specific data suite, including a large-scale, continual pre-training dataset with 129B tokens, an instruction tuning dataset with 214k samples, and a new medical evaluation benchmark (MIBE) across six critical medical tasks with 12 datasets. Our extensive evaluation using the MIBE shows that Me-LLaMA models achieve overall better performance than existing open-source medical LLMs in zero-shot, few-shot and supervised learning abilities. With task-specific instruction tuning, Me-LLaMA models outperform ChatGPT on 7 out of 8 datasets and GPT-4 on 5 out of 8 datasets. In addition, we investigated the catastrophic forgetting problem, and our results show that Me-LLaMA models outperform other open-source medical LLMs in mitigating this issue. Me-LLaMA is one of the largest open-source medical foundation LLMs that use both biomedical and clinical data. It exhibits superior performance across both general and medical tasks compared to other open-source medical LLMs, rendering it an attractive choice for medical AI applications. We release our models, datasets, and evaluation scripts at: https://github.com/BIDS-Xu-Lab/Me-LLaMA.

PubMed Disclaimer

Conflict of interest statement

COMPETING INTEREST The authors have no financial or non-financial conflicts of interest to disclose.

Figures

Figure 1.
Figure 1.
The comparison of the peak performance of Me-LLaMA models and Meditron 70B in the few-shot setting.
Figure 2.
Figure 2.
The comparison of performance of Me-LLaMA models in both zero-shot and task-specific instruction fine-tuning settings, against the zero-shot performance of ChatGPT and GPT-4.
Figure 3.
Figure 3.
Overview of the study. We first developed the Me-LLaMA base models by continual pre-training LLaMA2 models with mixed pre-training data (stage 1). Me-LLaMA-chat models are further developed by instruction fine-tuning Me-LLaMA base models (stage 2). We further fine-tuned Me-LLaMA base models with the task-specific training sets of evaluation datasets (stage 3) to evaluate their performance in the supervised learning setting and evaluate the performance of Me-LLaMA-chat models in the zero/few-shot setting (stage 4).

Similar articles

References

    1. Wei J. et al. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022).
    1. Bubeck S. et al. Sparks of artificial general intelligence: Early experiments with gpt-4. arXiv preprint arXiv:2303.12712 (2023).
    1. Nori H., King N., McKinney S. M., Carignan D. & Horvitz E. Capabilities of gpt-4 on medical challenge problems. arXiv preprint arXiv:2303.13375 (2023).
    1. Bengio Y., Ducharme R. & Vincent P. A neural probabilistic language model. Advances neural information processing systems 13 (2000).
    1. Brown T. et al. Language models are few-shot learners. Advances neural information processing systems 33, 1877–1901 (2020).

Publication types

LinkOut - more resources