Emotion Recognition Using EEG Signals and Audiovisual Features with Contrastive Learning
- PMID: 39451373
- PMCID: PMC11504283
- DOI: 10.3390/bioengineering11100997
Emotion Recognition Using EEG Signals and Audiovisual Features with Contrastive Learning
Abstract
Multimodal emotion recognition has emerged as a promising approach to capture the complex nature of human emotions by integrating information from various sources such as physiological signals, visual behavioral cues, and audio-visual content. However, current methods often struggle with effectively processing redundant or conflicting information across modalities and may overlook implicit inter-modal correlations. To address these challenges, this paper presents a novel multimodal emotion recognition framework which integrates audio-visual features with viewers' EEG data to enhance emotion classification accuracy. The proposed approach employs modality-specific encoders to extract spatiotemporal features, which are then aligned through contrastive learning to capture inter-modal relationships. Additionally, cross-modal attention mechanisms are incorporated for effective feature fusion across modalities. The framework, comprising pre-training, fine-tuning, and testing phases, is evaluated on multiple datasets of emotional responses. The experimental results demonstrate that the proposed multimodal approach, which combines audio-visual features with EEG data, is highly effective in recognizing emotions, highlighting its potential for advancing emotion recognition systems.
Keywords: contrastive learning; cross-attention mechanism; emotion recognition; multimodal learning.
Conflict of interest statement
The authors declare no conflicts of interest.
Figures




Similar articles
-
AVaTER: Fusing Audio, Visual, and Textual Modalities Using Cross-Modal Attention for Emotion Recognition.Sensors (Basel). 2024 Sep 10;24(18):5862. doi: 10.3390/s24185862. Sensors (Basel). 2024. PMID: 39338607 Free PMC article.
-
Cross-modal credibility modelling for EEG-based multimodal emotion recognition.J Neural Eng. 2024 Apr 11;21(2). doi: 10.1088/1741-2552/ad3987. J Neural Eng. 2024. PMID: 38565099
-
Research on cross-modal emotion recognition based on multi-layer semantic fusion.Math Biosci Eng. 2024 Jan 17;21(2):2488-2514. doi: 10.3934/mbe.2024110. Math Biosci Eng. 2024. PMID: 38454693
-
Emotion recognition in EEG signals using deep learning methods: A review.Comput Biol Med. 2023 Oct;165:107450. doi: 10.1016/j.compbiomed.2023.107450. Epub 2023 Sep 9. Comput Biol Med. 2023. PMID: 37708717 Review.
-
Multimodal fusion framework: a multiresolution approach for emotion classification and recognition from physiological signals.Neuroimage. 2014 Nov 15;102 Pt 1:162-72. doi: 10.1016/j.neuroimage.2013.11.007. Epub 2013 Nov 20. Neuroimage. 2014. PMID: 24269801 Review.
Cited by
-
The Effect of EEG Biofeedback Training Frequency and Environmental Conditions on Simple and Complex Reaction Times.Bioengineering (Basel). 2024 Nov 29;11(12):1208. doi: 10.3390/bioengineering11121208. Bioengineering (Basel). 2024. PMID: 39768026 Free PMC article.
-
Improved EEG-Based Emotion Classification via Stockwell Entropy and CSP Integration.Entropy (Basel). 2025 Apr 24;27(5):457. doi: 10.3390/e27050457. Entropy (Basel). 2025. PMID: 40422412 Free PMC article.
-
Cross-Subject Emotion Recognition with CT-ELCAN: Leveraging Cross-Modal Transformer and Enhanced Learning-Classify Adversarial Network.Bioengineering (Basel). 2025 May 15;12(5):528. doi: 10.3390/bioengineering12050528. Bioengineering (Basel). 2025. PMID: 40428147 Free PMC article.
References
-
- Andalibi N., Buss J. The human in emotion recognition on social media: Attitudes, outcomes, risks; Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems; Honolulu, HI, USA. 25–30 April 2020; pp. 1–16.
-
- Dubey A., Shingala B., Panara J.R., Desai K., Sahana M. Digital Content Recommendation System through Facial Emotion Recognition. Int. J. Res. Appl. Sci. Eng. Technol. 2023;11:1272–1276. doi: 10.22214/ijraset.2023.49225. - DOI
-
- Pepa L., Spalazzi L., Capecci M., Ceravolo M.G. Automatic emotion recognition in clinical scenario: A systematic review of methods. IEEE Trans. Affect. Comput. 2021;14:1675–1695. doi: 10.1109/TAFFC.2021.3128787. - DOI
-
- Caruelle D., Shams P., Gustafsson A., Lervik-Olsen L. Affective computing in marketing: Practical implications and research opportunities afforded by emotionally intelligent machines. Mark. Lett. 2022;33:163–169. doi: 10.1007/s11002-021-09609-0. - DOI
Grants and funding
LinkOut - more resources
Full Text Sources