Attentional Generative Multimodal Network for Neonatal Postoperative Pain Estimation
- PMID: 36939418
- PMCID: PMC10018439
- DOI: 10.1007/978-3-031-16437-8_72
Attentional Generative Multimodal Network for Neonatal Postoperative Pain Estimation
Abstract
Artificial Intelligence (AI)-based methods allow for automatic assessment of pain intensity based on continuous monitoring and processing of subtle changes in sensory signals, including facial expression, body movements, and crying frequency. Currently, there is a large and growing need for expanding current AI-based approaches to the assessment of postoperative pain in the neonatal intensive care unit (NICU). In contrast to acute procedural pain in the clinic, the NICU has neonates emerging from postoperative sedation, usually intubated, and with variable energy reserves for manifesting forceful pain responses. Here, we present a novel multi-modal approach designed, developed, and validated for assessment of neonatal postoperative pain in the challenging NICU setting. Our approach includes a robust network capable of efficient reconstruction of missing modalities (e.g., obscured facial expression due to intubation) using an unsupervised spatio-temporal feature learning with a generative model for learning the joint features. Our approach generates the final pain score along with the intensity using an attentional cross-modal feature fusion. Using experimental dataset from postoperative neonates in the NICU, our pain assessment approach achieves superior performance (AUC 0.906, accuracy 0.820) as compared to the state-of-the-art approaches.
Keywords: Generative model; Multimodal learning; NICU; Neonatal pain; Postoperative pain.
Figures
References
-
- Bowman S, Vilnis L, Vinyals O, Dai A, Jozefowicz R, Bengio S: Generating sentences from a continuous space. In: Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, pp. 10–21 (2016)
-
- Cao Q, Shen L, Xie W, Parkhi OM, Zisserman A: VGGFace2: a dataset for recognising faces across pose and age. In: 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), pp. 67–74. IEEE; (2018)
-
- Cao Y, Fleet DJ: Generalized product of experts for automatic and principled fusion of Gaussian process predictions. arXiv preprint arXiv:1410.7827 (2014)
-
- Choi JH, Lee JS: EmbraceNet: a robust deep learning architecture for multi-modal classification. Inf. Fusion 51, 259–270 (2019)
-
- Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L: ImageNet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255. IEEE; (2009)
Grants and funding
LinkOut - more resources
Full Text Sources