FusionSense: Emotion Classification Using Feature Fusion of Multimodal Data and Deep Learning in a Brain-Inspired Spiking Neural Network
- PMID: 32957655
- PMCID: PMC7571195
- DOI: 10.3390/s20185328
FusionSense: Emotion Classification Using Feature Fusion of Multimodal Data and Deep Learning in a Brain-Inspired Spiking Neural Network
Abstract
Using multimodal signals to solve the problem of emotion recognition is one of the emerging trends in affective computing. Several studies have utilized state of the art deep learning methods and combined physiological signals, such as the electrocardiogram (EEG), electroencephalogram (ECG), skin temperature, along with facial expressions, voice, posture to name a few, in order to classify emotions. Spiking neural networks (SNNs) represent the third generation of neural networks and employ biologically plausible models of neurons. SNNs have been shown to handle Spatio-temporal data, which is essentially the nature of the data encountered in emotion recognition problem, in an efficient manner. In this work, for the first time, we propose the application of SNNs in order to solve the emotion recognition problem with the multimodal dataset. Specifically, we use the NeuCube framework, which employs an evolving SNN architecture to classify emotional valence and evaluate the performance of our approach on the MAHNOB-HCI dataset. The multimodal data used in our work consists of facial expressions along with physiological signals such as ECG, skin temperature, skin conductance, respiration signal, mouth length, and pupil size. We perform classification under the Leave-One-Subject-Out (LOSO) cross-validation mode. Our results show that the proposed approach achieves an accuracy of 73.15% for classifying binary valence when applying feature-level fusion, which is comparable to other deep learning methods. We achieve this accuracy even without using EEG, which other deep learning methods have relied on to achieve this level of accuracy. In conclusion, we have demonstrated that the SNN can be successfully used for solving the emotion recognition problem with multimodal data and also provide directions for future research utilizing SNN for Affective computing. In addition to the good accuracy, the SNN recognition system is requires incrementally trainable on new data in an adaptive way. It only one pass training, which makes it suitable for practical and on-line applications. These features are not manifested in other methods for this problem.
Keywords: Evolving Spiking Neural Networks (eSNNs); NeuCube; Spatio-temporal data; facial emotion recognition; multimodal data.
Conflict of interest statement
The authors declare no conflict of interest.
Figures











Similar articles
-
NeuCube: a spiking neural network architecture for mapping, learning and understanding of spatio-temporal brain data.Neural Netw. 2014 Apr;52:62-76. doi: 10.1016/j.neunet.2014.01.006. Epub 2014 Jan 20. Neural Netw. 2014. PMID: 24508754 Review.
-
M1M2: Deep-Learning-Based Real-Time Emotion Recognition from Neural Activity.Sensors (Basel). 2022 Nov 3;22(21):8467. doi: 10.3390/s22218467. Sensors (Basel). 2022. PMID: 36366164 Free PMC article.
-
EEG-based emotion charting for Parkinson's disease patients using Convolutional Recurrent Neural Networks and cross dataset learning.Comput Biol Med. 2022 May;144:105327. doi: 10.1016/j.compbiomed.2022.105327. Epub 2022 Mar 11. Comput Biol Med. 2022. PMID: 35303579
-
Spatio-Temporal Representation of an Electoencephalogram for Emotion Recognition Using a Three-Dimensional Convolutional Neural Network.Sensors (Basel). 2020 Jun 20;20(12):3491. doi: 10.3390/s20123491. Sensors (Basel). 2020. PMID: 32575708 Free PMC article.
-
From Neural Networks to Emotional Networks: A Systematic Review of EEG-Based Emotion Recognition in Cognitive Neuroscience and Real-World Applications.Brain Sci. 2025 Feb 20;15(3):220. doi: 10.3390/brainsci15030220. Brain Sci. 2025. PMID: 40149742 Free PMC article. Review.
Cited by
-
Integrating Spatial and Temporal Information for Violent Activity Detection from Video Using Deep Spiking Neural Networks.Sensors (Basel). 2023 May 6;23(9):4532. doi: 10.3390/s23094532. Sensors (Basel). 2023. PMID: 37177737 Free PMC article.
-
Affective computing of multi-type urban public spaces to analyze emotional quality using ensemble learning-based classification of multi-sensor data.PLoS One. 2022 Jun 3;17(6):e0269176. doi: 10.1371/journal.pone.0269176. eCollection 2022. PLoS One. 2022. PMID: 35657805 Free PMC article.
-
A novel signal channel attention network for multi-modal emotion recognition.Front Neurorobot. 2024 Sep 11;18:1442080. doi: 10.3389/fnbot.2024.1442080. eCollection 2024. Front Neurorobot. 2024. PMID: 39323931 Free PMC article.
-
Domain adaptation spatial feature perception neural network for cross-subject EEG emotion recognition.Front Hum Neurosci. 2024 Dec 17;18:1471634. doi: 10.3389/fnhum.2024.1471634. eCollection 2024. Front Hum Neurosci. 2024. PMID: 39741785 Free PMC article.
-
CIT-EmotionNet: convolution interactive transformer network for EEG emotion recognition.PeerJ Comput Sci. 2024 Dec 23;10:e2610. doi: 10.7717/peerj-cs.2610. eCollection 2024. PeerJ Comput Sci. 2024. PMID: 39896395 Free PMC article.
References
-
- Calvo R.A., D’Mello S. Affect detection: An interdisciplinary review of models, methods, and their applications. IEEE Trans. Affect. Comput. 2010;1:18–37. doi: 10.1109/T-AFFC.2010.1. - DOI
-
- Fong T., Nourbakhsh I., Dautenhahn K. A survey of socially interactive robots. Robot. Auton. Syst. 2003;42:143–166. doi: 10.1016/S0921-8890(02)00372-X. - DOI
-
- Russell J.A. A circumplex model of affect. J. Personal. Soc. Psychol. 1980;39:1161. doi: 10.1037/h0077714. - DOI
-
- Gunes H., Schuller B., Pantic M., Cowie R. Emotion representation, analysis and synthesis in continuous space: A survey; Proceedings of the Face and Gesture 2011; Santa Barbara, CA, USA. 21–25 March 2011; pp. 827–834.
MeSH terms
LinkOut - more resources
Full Text Sources
Research Materials