Automatically Detecting Pain Using Facial Actions
- PMID: 21278824
- PMCID: PMC3296481
- DOI: 10.1109/ACII.2009.5349321
Automatically Detecting Pain Using Facial Actions
Abstract
Pain is generally measured by patient self-report, normally via verbal communication. However, if the patient is a child or has limited ability to communicate (i.e. the mute, mentally impaired, or patients having assisted breathing) self-report may not be a viable measurement. In addition, these self-report measures only relate to the maximum pain level experienced during a sequence so a frame-by-frame measure is currently not obtainable. Using image data from patients with rotator-cuff injuries, in this paper we describe an AAM-based automatic system which can detect pain on a frame-by-frame level. We do this two ways: directly (straight from the facial features); and indirectly (through the fusion of individual AU detectors). From our results, we show that the latter method achieves the optimal results as most discriminant features from each AU detector (i.e. shape or appearance) are used.
Figures
References
-
- Ashraf A, Lucey S, Cohn J, Chen T, Ambadar Z, Prkachin K, Solomon P, Theobald B-J. The painful face: pain expression recognition using active appearance models. Proceedings of the 9th international conference on Multimodal interfaces; Nagoya, Aichi: ACM; 2007. pp. 9–14.
-
- Brummer N, du Preez J. Application-independent evaluation of speaker detection. Computer Speech and Language. 2006;20:230–275.
-
- Cohn J, Ambadar Z, Ekman P. Observer-based measurement of facial measurement of facial expression with the facial action coding system. In: Coan A, Allen J, editors. The handbook of emotion elicitation and assessment. Oxford University Press; New York, USA: pp. 203–221.
-
- Cootes T, Edwards G, Taylor C. Active appearance models. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2001;23(6):681–685.
Grants and funding
LinkOut - more resources
Full Text Sources
Other Literature Sources