Capturing Requirements for a Data Annotation Tool for Intensive Care: Experimental User-Centered Design Study
- PMID: 39908549
- PMCID: PMC11840393
- DOI: 10.2196/56880
Capturing Requirements for a Data Annotation Tool for Intensive Care: Experimental User-Centered Design Study
Abstract
Background: Increasing use of computational methods in health care provides opportunities to address previously unsolvable problems. Machine learning techniques applied to routinely collected data can enhance clinical tools and improve patient outcomes, but their effective deployment comes with significant challenges. While some tasks can be addressed by training machine learning models directly on the collected data, more complex problems require additional input in the form of data annotations. Data annotation is a complex and time-consuming problem that requires domain expertise and frequently, technical proficiency. With clinicians' time being an extremely limited resource, existing tools fail to provide an effective workflow for deployment in health care.
Objective: This paper investigates the approach of intensive care unit staff to the task of data annotation. Specifically, it aims to (1) understand how clinicians approach data annotation and (2) capture the requirements for a digital annotation tool for the health care setting.
Methods: We conducted an experimental activity involving annotation of the printed excerpts of real time-series admission data with 7 intensive care unit clinicians. Each participant annotated an identical set of admissions with the periods of weaning from mechanical ventilation during a single 45-minute workshop. Participants were observed during task completion and their actions were analyzed within Norman's Interaction Cycle model to identify the software requirements.
Results: Clinicians followed a cyclic process of investigation, annotation, data reevaluation, and label refinement. Variety of techniques were used to investigate data and create annotations. We identified 11 requirements for the digital tool across 4 domains: annotation of individual admissions (n=5), semiautomated annotation (n=3), operational constraints (n=2), and use of labels in machine learning (n=1).
Conclusions: Effective data annotation in a clinical setting relies on flexibility in analysis and label creation and workflow continuity across multiple admissions. There is a need to ensure a seamless transition between data investigation, annotation, and refinement of the labels.
Keywords: ICU; annotation software; capturing software requirements; data annotation; data labeling; intensive care; machine learning.
©Marceli Wac, Raul Santos-Rodriguez, Chris McWilliams, Christopher Bourdeaux. Originally published in JMIR Human Factors (https://humanfactors.jmir.org), 05.02.2025.
Conflict of interest statement
Conflicts of Interest: None declared.
Figures




Similar articles
-
The future of Cochrane Neonatal.Early Hum Dev. 2020 Nov;150:105191. doi: 10.1016/j.earlhumdev.2020.105191. Epub 2020 Sep 12. Early Hum Dev. 2020. PMID: 33036834
-
DicomAnnotator: a Configurable Open-Source Software Program for Efficient DICOM Image Annotation.J Digit Imaging. 2020 Dec;33(6):1514-1526. doi: 10.1007/s10278-020-00370-w. J Digit Imaging. 2020. PMID: 32666365 Free PMC article.
-
A methodology for the annotation of surgical videos for supervised machine learning applications.Int J Comput Assist Radiol Surg. 2023 Sep;18(9):1673-1678. doi: 10.1007/s11548-023-02923-0. Epub 2023 May 28. Int J Comput Assist Radiol Surg. 2023. PMID: 37245179 Review.
-
Tell me something interesting: Clinical utility of machine learning prediction models in the ICU.J Biomed Inform. 2022 Aug;132:104107. doi: 10.1016/j.jbi.2022.104107. Epub 2022 Jun 7. J Biomed Inform. 2022. PMID: 35688332
-
Intelligent decision support with machine learning for efficient management of mechanical ventilation in the intensive care unit - A critical overview.Int J Med Inform. 2021 Jun;150:104469. doi: 10.1016/j.ijmedinf.2021.104469. Epub 2021 Apr 21. Int J Med Inform. 2021. PMID: 33906020 Review.
Cited by
-
Developing a Scalable Annotation Method for Large Datasets That Enhances Alarms With Actionability Data to Increase Informativeness: Mixed Methods Approach.J Med Internet Res. 2025 May 5;27:e65961. doi: 10.2196/65961. J Med Internet Res. 2025. PMID: 40324165 Free PMC article.
References
-
- Russell SJ, Norvig P. Prentice Hall Series in Artificial Intelligence. Singapore Taipei Tokyo: Pearson; 2016. Artificial Intelligence: A Modern Approach, Global Edition.
-
- Bellman R. An Introduction to Artificial Intelligence: Can Computers Think? San Francisco: Boyd & Fraser Publishing Company; 1978.
-
- Grosan C, Abraham A. Intelligent Systems Reference Library. Berlin: Springer-Verlag; 2011. Intelligent Systems: A Modern Approach.
-
- Hopfield JJ. Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA. 1982;79(8):2554–2558. doi: 10.1073/pnas.79.8.2554. https://europepmc.org/abstract/MED/6953413 - DOI - PMC - PubMed
-
- Introduction TML, Fourth E. Adaptive Computation and Machine Learning Series. Cambridge, MA: MIT Press; 2020. Introduction to Machine Learning, Fourth Edition.
MeSH terms
LinkOut - more resources
Full Text Sources
Medical