Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
- PMID: 40536801
- PMCID: PMC12226782
- DOI: 10.2196/70315
Large Language Model Architectures in Health Care: Scoping Review of Research Perspectives
Abstract
Background: Large language models (LLMs) can support health care professionals in their daily work, for example, when writing and filing reports or communicating diagnoses. With the rise of LLMs, current research investigates how LLMs could be applied in medical practice and their benefits for physicians in clinical workflows. However, most studies neglect the importance of selecting suitable LLM architectures.
Objective: In this literature review, we aim to provide insights on the different LLM model architecture families (ie, Bidirectional Encoder Representations from Transformers [BERT]-based or generative pretrained transformer [GPT]-based models) used in previous research. We report on the suitability and benefits of different LLM model architecture families for various research foci.
Methods: To this end, we conduct a scoping review to identify which LLMs are used in health care. Our search included manuscripts from PubMed, arXiv, and medRxiv. We used open and selective coding to assess the 114 identified manuscripts regarding 11 dimensions related to usage and technical facets and the research focus of the manuscripts.
Results: We identified 4 research foci that emerged previously in manuscripts, with LLM performance being the main focus. We found that GPT-based models are used for communicative purposes such as examination preparation or patient interaction. In contrast, BERT-based models are used for medical tasks such as knowledge discovery and model improvements.
Conclusions: Our study suggests that GPT-based models are better suited for communicative purposes such as report generation or patient interaction. BERT-based models seem to be better suited for innovative applications such as classification or knowledge discovery. This could be due to the architectural differences where GPT processes language unidirectionally and BERT bidirectionally, allowing more in-depth understanding of the text. In addition, BERT-based models seem to allow more straightforward extensions of their models for domain-specific tasks that generally lead to better results. In summary, health care professionals should consider the benefits and differences of the LLM architecture families when selecting a suitable model for their intended purpose.
Keywords: ChatGPT; digital health; generative artificial intelligence; large language models; medical informatics; scoping review.
©Florian Leiser, Richard Guse, Ali Sunyaev. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 19.06.2025.
Conflict of interest statement
Conflicts of Interest: None declared.
Figures
Similar articles
-
Implementing Large Language Models in Health Care: Clinician-Focused Review With Interactive Guideline.J Med Internet Res. 2025 Jul 11;27:e71916. doi: 10.2196/71916. J Med Internet Res. 2025. PMID: 40644686 Free PMC article. Review.
-
Use of Large Language Models to Classify Epidemiological Characteristics in Synthetic and Real-World Social Media Posts About Conjunctivitis Outbreaks: Infodemiology Study.J Med Internet Res. 2025 Jul 2;27:e65226. doi: 10.2196/65226. J Med Internet Res. 2025. PMID: 40601927 Free PMC article.
-
Applications and Concerns of ChatGPT and Other Conversational Large Language Models in Health Care: Systematic Review.J Med Internet Res. 2024 Nov 7;26:e22769. doi: 10.2196/22769. J Med Internet Res. 2024. PMID: 39509695 Free PMC article.
-
Applications of Large Language Models in the Field of Suicide Prevention: Scoping Review.J Med Internet Res. 2025 Jan 23;27:e63126. doi: 10.2196/63126. J Med Internet Res. 2025. PMID: 39847414 Free PMC article.
-
A dataset and benchmark for hospital course summarization with adapted large language models.J Am Med Inform Assoc. 2025 Mar 1;32(3):470-479. doi: 10.1093/jamia/ocae312. J Am Med Inform Assoc. 2025. PMID: 39786555
References
-
- OpenAI. Introducing ChatGPT. 2022. Nov 30, [2025-03-12]. https://openai.com/blog/chatgpt .
-
- Hu Krystal. ChatGPT sets record for fastest-growing user base—analyst note. 2023. Feb 01, [2025-05-09]. https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-u...
-
- Devlin J, Chang MW, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. 2019. May 24, [2025-04-18]. https://arxiv.org/abs/1810.04805 .
-
- Touvron H, Lavril T, Izacard G, Martinet X, Lachaux MA, Lacroix T, Rozière B, Goyal N, Hambro E, Azhar F, Rodriguez A, Joulin A, Grave E, Lample G. LLaMA: open and efficient foundation language models. ArXiv. posted online on 2023. 2023:1–27. https://arxiv.org/abs/2302.13971
-
- Kasneci E, Sessler K, Küchemann S, Bannert M, Dementieva D, Fischer F, Gasser U, Groh G, Günnemann S, Hüllermeier E, Krusche S, Kutyniok G, Michaeli T, Nerdel C, Pfeffer J, Poquet O, Sailer M, Schmidt A, Seidel T, Stadler M, Weller J, Kuhn J, Kasneci G. ChatGPT for good? On opportunities and challenges of large language models for education. Learn Indiv Differences. 2023;103(April 2023):1–9. doi: 10.1016/j.lindif.2023.102274. https://doi.org/10.1016/j.lindif.2023.102274 - DOI - DOI
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Medical