Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Jul 1;4(7):e2117391.
doi: 10.1001/jamanetworkopen.2021.17391.

Development and Validation of an Artificial Intelligence System to Optimize Clinician Review of Patient Records

Affiliations

Development and Validation of an Artificial Intelligence System to Optimize Clinician Review of Patient Records

Ethan Andrew Chi et al. JAMA Netw Open. .

Abstract

Importance: Physicians are required to work with rapidly growing amounts of medical data. Approximately 62% of time per patient is devoted to reviewing electronic health records (EHRs), with clinical data review being the most time-consuming portion.

Objective: To determine whether an artificial intelligence (AI) system developed to organize and display new patient referral records would improve a clinician's ability to extract patient information compared with the current standard of care.

Design, setting, and participants: In this prognostic study, an AI system was created to organize patient records and improve data retrieval. To evaluate the system on time and accuracy, a nonblinded, prospective study was conducted at a single academic medical center. Recruitment emails were sent to all physicians in the gastroenterology division, and 12 clinicians agreed to participate. Each of the clinicians participating in the study received 2 referral records: 1 AI-optimized patient record and 1 standard (non-AI-optimized) patient record. For each record, clinicians were asked 22 questions requiring them to search the assigned record for clinically relevant information. Clinicians reviewed records from June 1 to August 30, 2020.

Main outcomes and measures: The time required to answer each question, along with accuracy, was measured for both records, with and without AI optimization. Participants were asked to assess overall satisfaction with the AI system, their preferred review method (AI-optimized vs standard), and other topics to assess clinical utility.

Results: Twelve gastroenterology physicians/fellows completed the study. Compared with standard (non-AI-optimized) patient record review, the AI system saved first-time physician users 18% of the time used to answer the clinical questions (10.5 [95% CI, 8.5-12.6] vs 12.8 [95% CI, 9.4-16.2] minutes; P = .02). There was no significant decrease in accuracy when physicians retrieved important patient information (83.7% [95% CI, 79.3%-88.2%] with the AI-optimized vs 86.0% [95% CI, 81.8%-90.2%] without the AI-optimized record; P = .81). Survey responses from physicians were generally positive across all questions. Eleven of 12 physicians (92%) preferred the AI-optimized record review to standard review. Despite a learning curve pointed out by respondents, 11 of 12 physicians believed that the technology would save them time to assess new patient records and were interested in using this technology in their clinic.

Conclusions and relevance: In this prognostic study, an AI system helped physicians extract relevant patient information in a shorter time while maintaining high accuracy. This finding is particularly germane to the ever-increasing amounts of medical data and increased stressors on clinicians. Increased user familiarity with the AI system, along with further enhancements in the system itself, hold promise to further improve physician data extraction from large quantities of patient health records.

PubMed Disclaimer

Conflict of interest statement

Conflict of Interest Disclosures: None reported.

Figures

Figure 1.
Figure 1.. Data Flowchart and Artificial Intelligence (AI) System Pipeline
A, Forty-five readable patient referral records from gastroenterology clinicians were randomly obtained for the development of the system. They were separated into training, validation, and test sets. Three of the training set scans were manually annotated to identify information categories for the system to extract. Five records were used for physician evaluation: 4 for testing and 1 for collecting initial user feedback. B, The input is a scanned referral record, which is first partitioned into its constituent documents. The document is then classified into 1 of 9 categories (or undetermined): referral, note, laboratory, radiology, procedure, operative report, pathology, fax cover sheet, or insurance. The documents are ordered by their most recent date. Laboratory values are extracted and presented in a table, sorted by date. EHR indicates electronic health record.
Figure 2.
Figure 2.. User Interface With Artificial Intelligence (AI)–Optimized Record
A, The original referral record PDF is displayed on the right of the interface. The AI output is shown on the left in 3 representative sections: B, the social information section, which contains smoking and allergy information; C, the radiology section, which displays items that are predicted to belong to the radiology category; and D, the laboratory table, which organizes by date the laboratory values extracted from the document. In practice, 1 section for each category would be predicted by the system to be in the document.
Figure 3.
Figure 3.. Study Design and Record Assignment for Study
A, Clinicians were assigned 1 standard and 1 artificial intelligence (AI)–optimized record in random order. B, Combination of records assigned to each participant.
Figure 4.
Figure 4.. Time Saved by Artificial Intelligence (AI) Optimization
A, Per-physician time taken for completion of questions for AI-optimized and standard review. Bars with an orange dot on the left and a blue dot on the right represent time saved with AI optimization; bars with a blue dot on the left and an orange dot on the right represent time lost. B, Time saved, adjusted for record size. The left side shows each physician’s time saved using AI-optimized review; the right side shows their time saved after using the time savings mixed model to standardize with a record size of 34 pages. Boxes represent quartiles (Q3-Q1), horizontal bars represent the median, and orange diamonds represent the mean. Error bars indicate the maximum and minimum values that are not outliers. Blue circles represent the time saved for each individual person. C, Association between standard review time and time saved with AI optimization. There was a correlation between the time physicians take to complete a standard review and the time saved with AI optimization (r = 0.80; P = .002).
Figure 5.
Figure 5.. Subjective Feedback Results
After the 2 evaluation referral records, a subjective feedback survey was administered. Feedback from clinicians on software utility and performance on a 5-point Likert scale was generally positive. For any given survey question or statement, the absence of a percentage of respondents indicates that there were no responses for that category. AI indicates artificial intelligence.

Comment in

References

    1. Myrick KL, Ogburn DF, Ward BW. Table. Percentage of office-based physicians using any electronic health record (EHR)/electronic medical record (EMR) system and physicians that have a certified EHR/EMR system, by US state: National Electronic Health Records Survey, 2017. January 2019. Accessed November 2020. https://www.cdc.gov/nchs/data/nehrs/2017_NEHRS_Web_Table_EHR_State.pdf
    1. Sinsky C, Colligan L, Li L, et al. . Allocation of physician time in ambulatory practice: a time and motion study in 4 specialties. Ann Intern Med. 2016;165(11):753-760. doi:10.7326/M16-0961 - DOI - PubMed
    1. Datta S, Bernstam EV, Roberts K. A frame semantic overview of NLP-based information extraction for cancer-related EHR notes. J Biomed Inform. 2019;100:103301. doi:10.1016/j.jbi.2019.103301 - DOI - PMC - PubMed
    1. Ohno-Machado L, Nadkarni P, Johnson K. Natural language processing: algorithms and tools to extract computable information from EHRs and from the biomedical literature. J Am Med Inform Assoc. 2013;20(5):805. doi:10.1136/amiajnl-2013-002214 - DOI - PMC - PubMed
    1. Arndt BG, Beasley JW, Watkinson MD, et al. . Tethered to the EHR: primary care physician workload assessment using EHR event log data and time-motion observations. Ann Fam Med. 2017;15(5):419-426. doi:10.1370/afm.2121 - DOI - PMC - PubMed