Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Mar 24;14(7):2213.
doi: 10.3390/jcm14072213.

Preserving Informative Presence: How Missing Data and Imputation Strategies Affect the Performance of an AI-Based Early Warning Score

Affiliations

Preserving Informative Presence: How Missing Data and Imputation Strategies Affect the Performance of an AI-Based Early Warning Score

Taeyong Sim et al. J Clin Med. .

Abstract

Background/Objectives: Data availability can affect the performance of AI-based early warning scores (EWSs). This study evaluated how the extent of missing data and imputation strategies influence the predictive performance of the VitalCare-Major Adverse Event Score (VC-MAES), an AI-based EWS that uses last observation carried forward and normal-value imputation for missing values, to forecast clinical deterioration events, including unplanned ICU transfers, cardiac arrests, or death, up to 6 h in advance. Methods: We analyzed real-world data from 6039 patient encounters at Keimyung University Dongsan Hospital, Republic of Korea. Performance was evaluated under three scenarios: (1) using only vital signs and age, treating all other variables as missing; (2) reintroducing a full set of real-world clinical variables; and (3) imputing missing values drawn from a distribution within one standard deviation of the observed mean or using Multiple Imputation by Chained Equations (MICE). Results: VC-MAES achieved the area under the receiver operating characteristic curve (AUROC) of 0.896 using only vital signs and age, outperforming traditional EWSs, including the National Early Warning Score (0.797) and the Modified Early Warning Score (0.722). Reintroducing full clinical variables improved the AUROC to 0.918, whereas mean-based imputation or MICE decreased the performance to 0.885 and 0.827, respectively. Conclusions: VC-MAES demonstrates robust predictive performance with limited inputs, outperforming traditional EWSs. Incorporating actual clinical data significantly improved accuracy. In contrast, mean-based or MICE imputation yielded poorer results than the default normal-value imputation, potentially due to disregarding the "informative presence" embedded in missing data patterns. These findings underscore the importance of understanding missingness patterns and employing imputation strategies that consider the decision-making context behind data availability to enhance model reliability.

Keywords: artificial intelligence; early warning score; imputation; modified early warning score; national early warning score.

PubMed Disclaimer

Conflict of interest statement

Authors Taeyong Sim, Sangchul Hahn, Kwang-Joon Kim, Eun-Young Cho, Yeeun Jeong, Ji-hyun Kim and Ki-Byung Lee were employed by the company AITRICS. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

Figure 1
Figure 1
Receiver Operating Characteristic Curves (a) and Precision–Recall Curves (b) comparing predictive performance among models using vital signs only, vital signs + laboratory data, Modified Early Warning Score (MEWS), and National Early Warning Score (NEWS).
Figure 2
Figure 2
Receiver Operating Characteristic Curves (a) and Precision–Recall Curves (b) comparing predictive performance among models using vital signs + laboratory data and forced lab imputation.

Similar articles

Cited by

References

    1. Liu F., Panagiotakos D. Real-world data: A brief review of the methods, applications, challenges and opportunities. BMC Med. Res. Methodol. 2022;22:287. doi: 10.1186/s12874-022-01768-6. - DOI - PMC - PubMed
    1. Mendelsohn A.B., Dreyer N.A., Mattox P.W., Su Z., Swenson A., Li R., Turner J.R., Velentgas P. Characterization of Missing Data in Clinical Registry Studies. Ther. Innov. Regul. Sci. 2015;49:146–154. doi: 10.1177/2168479014532259. - DOI - PubMed
    1. Goldstein B.A., Navar A.M., Pencina M.J., Ioannidis J.P.A. Opportunities and challenges in developing risk prediction models with electronic health records data: A systematic review. J. Am. Med. Inform. Assoc. 2017;24:198–208. doi: 10.1093/jamia/ocw042. - DOI - PMC - PubMed
    1. Getzen E., Ungar L., Mowery D., Jiang X., Long Q. Mining for equitable health: Assessing the impact of missing data in electronic health records. J. Biomed. Inform. 2023;139:104269. doi: 10.1016/j.jbi.2022.104269. - DOI - PMC - PubMed
    1. Ayilara O.F., Zhang L., Sajobi T.T., Sawatzky R., Bohm E., Lix L.M. Impact of missing data on bias and precision when estimating change in patient-reported outcomes from a clinical registry. Health Qual. Life Outcomes. 2019;17:106. doi: 10.1186/s12955-019-1181-2. - DOI - PMC - PubMed

LinkOut - more resources