Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 May;28(5):465-469.
doi: 10.1007/s10157-023-02451-w. Epub 2024 Feb 14.

Performance of ChatGPT and Bard in self-assessment questions for nephrology board renewal

Affiliations

Performance of ChatGPT and Bard in self-assessment questions for nephrology board renewal

Ryunosuke Noda et al. Clin Exp Nephrol. 2024 May.

Abstract

Background: Large language models (LLMs) have impacted advances in artificial intelligence. While LLMs have demonstrated high performance in general medical examinations, their performance in specialized areas such as nephrology is unclear. This study aimed to evaluate ChatGPT and Bard in their potential nephrology applications.

Methods: Ninety-nine questions from the Self-Assessment Questions for Nephrology Board Renewal from 2018 to 2022 were presented to two versions of ChatGPT (GPT-3.5 and GPT-4) and Bard. We calculated the correct answer rates for the five years, each year, and question categories and checked whether they exceeded the pass criterion. The correct answer rates were compared with those of the nephrology residents.

Results: The overall correct answer rates for GPT-3.5, GPT-4, and Bard were 31.3% (31/99), 54.5% (54/99), and 32.3% (32/99), respectively, thus GPT-4 significantly outperformed GPT-3.5 (p < 0.01) and Bard (p < 0.01). GPT-4 passed in three years, barely meeting the minimum threshold in two. GPT-4 demonstrated significantly higher performance in problem-solving, clinical, and non-image questions than GPT-3.5 and Bard. GPT-4's performance was between third- and fourth-year nephrology residents.

Conclusions: GPT-4 outperformed GPT-3.5 and Bard and met the Nephrology Board renewal standards in specific years, albeit marginally. These results highlight LLMs' potential and limitations in nephrology. As LLMs advance, nephrologists should understand their performance for future applications.

Keywords: Artificial intelligence; ChatGPT; GPT-4; Large language models; Nephrology.

PubMed Disclaimer

References

    1. Zhao WX, Zhou K, Li J et al. A survey of large language models. ArXiv e-prints, 2023 ( arXiv:2303.18223 ).
    1. Gilson A, Safranek CW, Huang T, et al. How does ChatGPT perform on the united states medical licensing examination? the implications of large language models for medical education and knowledge assessment. JMIR Med Educ. 2023;9: e45312. - DOI - PubMed - PMC
    1. Kung TH, Cheatham M, Medenilla A, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health. 2023;2: e0000198. - DOI - PubMed - PMC
    1. Sallam M. The utility of ChatGPT as an example of large language models in healthcare education, research and practice: systematic review on the future perspectives and potential limitations. MedRxiv e-prints, 2023 (medRxiv: 2023.02.19.23286155v1).
    1. Introducing ChatGPT: OpenAI. https://openai.com/blog/chatgpt/ . Published November 30, 2022. Accessed 25 May 25 2023.

LinkOut - more resources