Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Mar 17;17(3):e80737.
doi: 10.7759/cureus.80737. eCollection 2025 Mar.

Evaluating the Performance and Safety of Large Language Models in Generating Type 2 Diabetes Mellitus Management Plans: A Comparative Study With Physicians Using Real Patient Records

Affiliations

Evaluating the Performance and Safety of Large Language Models in Generating Type 2 Diabetes Mellitus Management Plans: A Comparative Study With Physicians Using Real Patient Records

Agnibho Mondal et al. Cureus. .

Abstract

Background The integration of large language models (LLMs) such as GPT-4 into healthcare presents potential benefits and challenges. While LLMs show promise in applications ranging from scientific writing to personalized medicine, their practical utility and safety in clinical settings remain under scrutiny. Concerns about accuracy, ethical considerations, and bias necessitate rigorous evaluation of these technologies against established medical standards. Methods This study involved a comparative analysis using anonymized patient records from a healthcare setting in the state of West Bengal, India. Management plans for 50 patients with type 2 diabetes mellitus were generated by GPT-4 and three physicians, who were blinded to each other's responses. These plans were evaluated against a reference management plan based on American Diabetes Society guidelines. Completeness, necessity, and dosage accuracy were quantified and a Prescribing Error Score was devised to assess the quality of the generated management plans. The safety of the management plans generated by GPT-4 was also assessed. Results Results indicated that physicians' management plans had fewer missing medications compared to those generated by GPT-4 (p=0.008). However, GPT-4-generated management plans included fewer unnecessary medications (p=0.003). No significant difference was observed in the accuracy of drug dosages (p=0.975). The overall error scores were comparable between physicians and GPT-4 (p=0.301). Safety issues were noted in 16% of the plans generated by GPT-4, highlighting potential risks associated with AI-generated management plans. Conclusion The study demonstrates that while GPT-4 can effectively reduce unnecessary drug prescriptions, it does not yet match the performance of physicians in terms of plan completeness. The findings support the use of LLMs as supplementary tools in healthcare, highlighting the need for enhanced algorithms and continuous human oversight to ensure the efficacy and safety of artificial intelligence in clinical settings.

Keywords: artificial intelligence in medicine; clinical decision support system; diabetes mellitus type 2; large language model(llm); oral hypoglycemic agents.

PubMed Disclaimer

Conflict of interest statement

Human subjects: Consent for treatment and open access publication was obtained or waived by all participants in this study. Clinimed Independent Ethics Committee, Kolkata issued approval CLPL/CIEC/001/2024 dated May 11, 2024. Animal subjects: All authors have confirmed that this study did not involve animal subjects or tissue. Conflicts of interest: In compliance with the ICMJE uniform disclosure form, all authors declare the following: Payment/services info: All authors have declared that no financial support was received from any organization for the submitted work. Financial relationships: All authors have declared that they have no financial relationships at present or within the previous three years with any organizations that might have an interest in the submitted work. Other relationships: All authors have declared that there are no other relationships or activities that could appear to have influenced the submitted work.

Figures

Figure 1
Figure 1. Methodological Framework for Comparing Physician and AI-Generated Diabetes Management Plans
Figure 2
Figure 2. Comparison of Error Scores: Physicians vs. GPT-4 in Type 2 Diabetes Management

Similar articles

References

    1. Assessing the research landscape and clinical utility of large language models: a scoping review. Park YJ, Pillai A, Deng J, Guo E, Gupta M, Paget M, Naugler C. BMC Med Inform Decis Mak. 2024;24:72. - PMC - PubMed
    1. GPT-4 technical report [PREPRINT] OpenAI OpenAI, Achiam J, Adler S, et al. arXiv. 2023
    1. Testing and evaluation of health care applications of large language models: a systematic review. Bedi S, Liu Y, Orr-Ewing L, et al. JAMA. 2025;333:319–328. - PMC - PubMed
    1. Evaluating large language models as agents in the clinic. Mehandru N, Miao BY, Almaraz ER, Sushil M, Butte AJ, Alaa A. NPJ Digit Med. 2024;7:84. - PMC - PubMed
    1. A pilot study on the efficacy of GPT-4 in providing orthopedic treatment recommendations from MRI reports. Truhn D, Weber CD, Braun BJ, Bressem K, Kather JN, Kuhl C, Nebelung S. Sci Rep. 2023;13:20159. - PMC - PubMed

LinkOut - more resources