Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Aug 23;5(1):100600.
doi: 10.1016/j.xops.2024.100600. eCollection 2025 Jan-Feb.

Analysis of ChatGPT Responses to Ophthalmic Cases: Can ChatGPT Think like an Ophthalmologist?

Affiliations

Analysis of ChatGPT Responses to Ophthalmic Cases: Can ChatGPT Think like an Ophthalmologist?

Jimmy S Chen et al. Ophthalmol Sci. .

Abstract

Objective: Large language models such as ChatGPT have demonstrated significant potential in question-answering within ophthalmology, but there is a paucity of literature evaluating its ability to generate clinical assessments and discussions. The objectives of this study were to (1) assess the accuracy of assessment and plans generated by ChatGPT and (2) evaluate ophthalmologists' abilities to distinguish between responses generated by clinicians versus ChatGPT.

Design: Cross-sectional mixed-methods study.

Subjects: Sixteen ophthalmologists from a single academic center, of which 10 were board-eligible and 6 were board-certified, were recruited to participate in this study.

Methods: Prompt engineering was used to ensure ChatGPT output discussions in the style of the ophthalmologist author of the Medical College of Wisconsin Ophthalmic Case Studies. Cases where ChatGPT accurately identified the primary diagnoses were included and then paired. Masked human-generated and ChatGPT-generated discussions were sent to participating ophthalmologists to identify the author of the discussions. Response confidence was assessed using a 5-point Likert scale score, and subjective feedback was manually reviewed.

Main outcome measures: Accuracy of ophthalmologist identification of discussion author, as well as subjective perceptions of human-generated versus ChatGPT-generated discussions.

Results: Overall, ChatGPT correctly identified the primary diagnosis in 15 of 17 (88.2%) cases. Two cases were excluded from the paired comparison due to hallucinations or fabrications of nonuser-provided data. Ophthalmologists correctly identified the author in 77.9% ± 26.6% of the 13 included cases, with a mean Likert scale confidence rating of 3.6 ± 1.0. No significant differences in performance or confidence were found between board-certified and board-eligible ophthalmologists. Subjectively, ophthalmologists found that discussions written by ChatGPT tended to have more generic responses, irrelevant information, hallucinated more frequently, and had distinct syntactic patterns (all P < 0.01).

Conclusions: Large language models have the potential to synthesize clinical data and generate ophthalmic discussions. While these findings have exciting implications for artificial intelligence-assisted health care delivery, more rigorous real-world evaluation of these models is necessary before clinical deployment.

Financial disclosures: The author(s) have no proprietary or commercial interest in any materials discussed in this article.

Keywords: Artificial Intelligence; ChatGPT; Large language models; Medical Education; Ophthalmology.

PubMed Disclaimer

Figures

Figure 1
Figure 1
Percentage of accurate identification of ChatGPT vs. ophthalmologist discussion authorship by board-eligible and board-certified ophthalmologists.
Figure 2
Figure 2
Correlation between Likert scale confidence score and response accuracy regarding authorship by respondent.

Similar articles

Cited by

References

    1. Gulshan V., Peng L., Coram M., et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016;316(22):2402–2410. - PubMed
    1. Brown J.M., Campbell J.P., Beers A., et al. For the imaging and informatics in retinopathy of prematurity (i-ROP) research consortium. Automated diagnosis of plus disease in retinopathy of prematurity using deep convolutional neural networks. JAMA Ophthalmol. 2018;136(7):803–810. - PMC - PubMed
    1. Baxter S.L., Klie A.R., Radha S.B., et al. Text processing for detection of fungal ocular involvement in critical care patients: cross-sectional study. J Med Internet Res. 2020;22(8) - PMC - PubMed
    1. Applications of natural language processing in ophthalmology: present and future - PMC. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9393550/ - PMC - PubMed
    1. Chen J.S., Coyner A.S., Chan R.V.P., et al. Deepfakes in ophthalmology: applications and realism of synthetic retinal images from generative adversarial networks. Ophthalmol Sci. 2021;1(4) - PMC - PubMed