Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Jul 25;13(1):11990.
doi: 10.1038/s41598-023-38868-2.

Speech emotion classification using attention based network and regularized feature selection

Affiliations

Speech emotion classification using attention based network and regularized feature selection

Samson Akinpelu et al. Sci Rep. .

Abstract

Speech emotion classification (SEC) has gained the utmost height and occupied a conspicuous position within the research community in recent times. Its vital role in Human-Computer Interaction (HCI) and affective computing cannot be overemphasized. Many primitive algorithmic solutions and deep neural network (DNN) models have been proposed for efficient recognition of emotion from speech however, the suitability of these methods to accurately classify emotion from speech with multi-lingual background and other factors that impede efficient classification of emotion is still demanding critical consideration. This study proposed an attention-based network with a pre-trained convolutional neural network and regularized neighbourhood component analysis (RNCA) feature selection techniques for improved classification of speech emotion. The attention model has proven to be successful in many sequence-based and time-series tasks. An extensive experiment was carried out using three major classifiers (SVM, MLP and Random Forest) on a publicly available TESS (Toronto English Speech Sentence) dataset. The result of our proposed model (Attention-based DCNN+RNCA+RF) achieved 97.8% classification accuracy and yielded a 3.27% improved performance, which outperforms state-of-the-art SEC approaches. Our model evaluation revealed the consistency of attention mechanism and feature selection with human behavioural patterns in classifying emotion from auditory speech.

PubMed Disclaimer

Conflict of interest statement

The authors declare no competing interests.

Figures

Figure 1
Figure 1
Conventional speech emotion classification system.
Figure 2
Figure 2
Structure of mel-spectrogram extraction.
Figure 3
Figure 3
Proposed model architecture.
Figure 4
Figure 4
Convolutional layers block diagram.
Figure 5
Figure 5
RNCA framework.
Figure 6
Figure 6
Attention-based Vgg16+RNCA+RF.
Figure 7
Figure 7
Attention-based Vgg16+RNCA+MLP.
Figure 8
Figure 8
Attention-based Vgg16+RNCA+SVM.
Figure 9
Figure 9
Attention-based Vgg19+RNCA+RF.
Figure 10
Figure 10
Attention-based Vgg19+RNCA+MLP.
Figure 11
Figure 11
Attention-based Vgg19+RNCA+SVM.
Figure 12
Figure 12
Performance chart with 4 metrics and 3 classifiers.
Figure 13
Figure 13
Model loss curve.
Figure 14
Figure 14
ROC curve.

Similar articles

Cited by

References

    1. Costantini G, Parada-Cabaleiro E, Casali D, Cesarini V. The emotion probe: On the universality of cross-linguistic and cross-gender speech emotion recognition via machine learning. Sensors. 2022 doi: 10.3390/s22072461. - DOI - PMC - PubMed
    1. Chimthankar, P. P. Speech Emotion Recognition using Deep Learning. http://norma.ncirl.ie/5142/1/priyankaprashantchimthankar.pdf (2021)
    1. Saad, H. F.and Mahmud, Shaheen, M., Hasan, M., Farastu, P. & Kabir, M. Is speech emotion recognition language-independent? Analysis of english and bangla languages using language-independent vocal features. arXiv:2111.10776 (2021)
    1. Burghardt GM. A place for emotions in behavior systems research. Behavioural Process. 2019 doi: 10.1016/j.beproc.2019.06.004. - DOI - PubMed
    1. Mustaqeem, Kwon S. The emotion probe: On the universality of cross-linguistic and cross-gender speech emotion recognition via machine learning. Appl. Soft Comput. 2021 doi: 10.1016/j.asoc.2021.107101. - DOI - PMC - PubMed