Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 May 1;5(5):e0897.
doi: 10.1097/CCE.0000000000000897. eCollection 2023 May.

Developing and Validating a Prediction Model For Death or Critical Illness in Hospitalized Adults, an Opportunity for Human-Computer Collaboration

Affiliations

Developing and Validating a Prediction Model For Death or Critical Illness in Hospitalized Adults, an Opportunity for Human-Computer Collaboration

Amol A Verma et al. Crit Care Explor. .

Abstract

Hospital early warning systems that use machine learning (ML) to predict clinical deterioration are increasingly being used to aid clinical decision-making. However, it is not known how ML predictions complement physician and nurse judgment. Our objective was to train and validate a ML model to predict patient deterioration and compare model predictions with real-world physician and nurse predictions.

Design: Retrospective and prospective cohort study.

Setting: Academic tertiary care hospital.

Patients: Adult general internal medicine hospitalizations.

Measurements and main results: We developed and validated a neural network model to predict in-hospital death and ICU admission in 23,528 hospitalizations between April 2011 and April 2019. We then compared model predictions with 3,374 prospectively collected predictions from nurses, residents, and attending physicians about their own patients in 960 hospitalizations between April 30, and August 28, 2019. ML model predictions achieved clinician-level accuracy for predicting ICU admission or death (ML median F1 score 0.32 [interquartile range (IQR) 0.30-0.34], AUC 0.77 [IQ 0.76-0.78]; clinicians median F1-score 0.33 [IQR 0.30-0.35], AUC 0.64 [IQR 0.63-0.66]). ML predictions were more accurate than clinicians for ICU admission. Of all ICU admissions and deaths, 36% occurred in hospitalizations where the model and clinicians disagreed. Combining human and model predictions detected 49% of clinical deterioration events, improving sensitivity by 16% compared with clinicians alone and 24% compared with the model alone while maintaining a positive predictive value of 33%, thus keeping false alarms at a clinically acceptable level.

Conclusions: ML models can complement clinician judgment to predict clinical deterioration in hospital. These findings demonstrate important opportunities for human-computer collaboration to improve prognostication and personalized medicine in hospital.

Keywords: artificial intelligence; clinical prediction; machine learning; mortality; prognosis.

PubMed Disclaimer

Figures

Figure 1.
Figure 1.
Accuracy of clinicians and the machine learning (ML) model in predicting ICU and death across 500 bootstrapped samples. Boxplots depict the distribution (horizontal bar denotes median, box is 25/75 percentile) of the F1-score of the ML model and clinicians across 500 bootstrapped samples. Different bootstrapped samples were used for each clinician type, resulting in slightly different estimates of ML model performance in each comparison. The ML model demonstrates superior F1-scores in predicting ICU admission, but the model was not more accurate than clinicians in predicting death. Combining model and clinician predictions leads to improvements in F1-scores, which is particularly important given that the value of identifying true positive cases outweighs identification of true negatives in an early warning system.

References

    1. Van Galen LS, Struik PW, Driesen BEJM, et al. : Delayed recognition of deterioration of patients in general wards is mostly caused by human related monitoring failures: A root cause analysis of unplanned ICU admissions. PLoS One] 2016; 11:e01613931–e01613914 - PMC - PubMed
    1. Detsky ME, Harhay MO, Bayard DF, et al. : Discriminative accuracy of physician and nurse predictions for survival and functional outcomes 6 months after an ICU admission. JAMA 2017; 317:2187–2195 - PMC - PubMed
    1. Rojas JC, Lyons PG, Jiang T, et al. : Accuracy of clinicians’ ability to predict the need for intensive care unit readmission. Ann Am Thorac Soc 2020; 17:847–853 - PMC - PubMed
    1. Sinuff T, Adhikari NKJ, Cook DJ, et al. : Mortality predictions in the intensive care unit: Comparing physicians with scoring systems*. Crit Care Med 2006; 34878–885 - PubMed
    1. van Doorn WPTM, Stassen PM, Borggreve HF, et al. : A comparison of machine learning models versus clinical evaluation for mortality prediction in patients with sepsis. PLoS One 2021; 16:e02451571–e02451515 - PMC - PubMed