Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Comparative Study
. 2025 Aug;35(8):4549-4557.
doi: 10.1007/s00330-025-11438-y. Epub 2025 Feb 20.

Large language models for error detection in radiology reports: a comparative analysis between closed-source and privacy-compliant open-source models

Affiliations
Comparative Study

Large language models for error detection in radiology reports: a comparative analysis between closed-source and privacy-compliant open-source models

Babak Salam et al. Eur Radiol. 2025 Aug.

Abstract

Purpose: Large language models (LLMs) like Generative Pre-trained Transformer 4 (GPT-4) can assist in detecting errors in radiology reports, but privacy concerns limit their clinical applicability. This study compares closed-source and privacy-compliant open-source LLMs for detecting common errors in radiology reports.

Materials and methods: A total of 120 radiology reports were compiled (30 each from X-ray, ultrasound, CT, and MRI). Subsequently, 397 errors from five categories (typographical, numerical, findings-impression discrepancies, omission/insertion, interpretation) were inserted into 100 of these reports; 20 reports were left unchanged. Two open-source models (Llama 3-70b, Mixtral 8x22b) and two commercial closed-source (GPT-4, GPT-4o) were tasked with error detection using identical prompts. The Kruskall-Wallis test and paired t-test were used for statistical analysis.

Results: Open-source LLMs required less processing time per radiology report than closed-source LLMs (6 ± 2 s vs. 13 ± 4 s; p < 0.001). Closed-source LLMs achieved higher error detection rates than open-source LLMs (GPT-4o: 88% [348/397; 95% CI: 86, 92], GPT-4: 83% [328/397; 95% CI: 80, 87], Llama 3-70b: 79% [311/397; 95% CI: 76, 83], Mixtral 8x22b: 73% [288/397; 95% CI: 68, 77]; p < 0.001). Numerical errors (88% [67/76; 95% CI: 82, 93]) were detected significantly more often than typographical errors (75% [65/86; 95% CI: 68, 82]; p = 0.02), discrepancies between findings and impression (73% [73/101; 95% CI: 67, 80]; p < 0.01), and interpretation errors (70% [50/71; 95% CI: 62, 78]; p = 0.001).

Conclusion: Open-source LLMs demonstrated effective error detection, albeit with comparatively lower accuracy than commercial closed-source models, and have potential for clinical applications when deployed via privacy-compliant local hosting solutions.

Key points: Question Can privacy-compliant open-source large language models (LLMs) match the error-detection performance of commercial non-privacy-compliant closed-source models in radiology reports? Findings Closed-source LLMs achieved slightly higher accuracy in detecting radiology report errors than open-source models, with Llama 3-70b yielding the best results among the open-source models. Clinical relevance Open-source LLMs offer a privacy-compliant alternative for automated error detection in radiology reports, improving clinical workflow efficiency while ensuring patient data confidentiality. Further refinement could enhance their accuracy, contributing to better diagnosis and patient care.

Keywords: Error detection; Generative pre-trained transformers; Large language models; Open source; Radiology reports.

PubMed Disclaimer

Conflict of interest statement

Compliance with ethical standards. Guarantor: The scientific guarantor of this publication is Alexander Isaak, MD. Conflict of interest: The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article. Statistics and biometry: No complex statistical methods were necessary for this paper. Informed consent: Informed consent was not required as no patient-identifying information was used. Ethical approval: After consultation, specific approval from the institutional review board was not required as no patient-identifying information was used. Study subjects or cohorts overlap: Not applicable. Methodology: Experimental

Figures

Fig. 1
Fig. 1
Study flowchart. A total of 120 radiology reports were compiled, with 30 reports from each modality (X-ray, ultrasound, CT, and MRI). Errors were then intentionally inserted into 100 of these reports, while 20 reports remained unchanged. Two commercial closed-source (GPT-4, GPT-4o) and two open-source models (Llama 3-70b, Mixtral 8x22b) were tasked with detecting errors using identical prompts
Fig. 2
Fig. 2
Example of a radiology report with highlighted errors, along with the corresponding correction output generated by GPT-4o
Fig. 3
Fig. 3
Proofreading example with different detection performance of GPT-4, GPT-4o, Llama 3-70b, and Mixtral 8x22b. Integrated errors are highlighted in red in the text. For each model, correct detection of the error is marked with a green tick, non-detection of an error with a red cross
Fig. 4
Fig. 4
Bar graph shows the percentage of detected errors for GPT-4o, GPT-4, Llama 3-70b, and Mixtral 8x22b. The error bars are 95% CIs

References

    1. McDonald RJ, Schwartz KM, Eckel LJ et al (2015) The effects of changes in utilization and technological advancements of cross-sectional imaging on radiologist workload. Acad Radiol 22:1191–1198 - PubMed
    1. Zhou L, Blackley SV, Kowalski L et al (2018) Analysis of errors in dictated clinical documents assisted by speech recognition software and professional transcriptionists. JAMA Netw Open 1:e180530 - PMC - PubMed
    1. Quint LE, Quint DJ, Myles JD (2008) Frequency and spectrum of errors in final radiology reports generated with automatic speech recognition technology. J Am Coll Radiol 5:1196–1199 - PubMed
    1. Onder O, Yarasir Y, Azizova A, Durhan G, Onur MR, Ariyurek OM (2021) Errors, discrepancies and underlying bias in radiology with case examples: a pictorial review. Insights Imaging 12:51 - PMC - PubMed
    1. Salam B, Kravchenko D, Nowak S et al (2024) Generative pre-trained transformer 4 makes cardiovascular magnetic resonance reports easy to understand. J Cardiovasc Magn Reson 26:101035 - PMC - PubMed

Publication types

LinkOut - more resources