Robust EEG-Based Decoding of Auditory Attention With High-RMS-Level Speech Segments in Noisy Conditions
- PMID: 33132874
- PMCID: PMC7576187
- DOI: 10.3389/fnhum.2020.557534
Robust EEG-Based Decoding of Auditory Attention With High-RMS-Level Speech Segments in Noisy Conditions
Abstract
The attended speech stream can be detected robustly, even in adverse auditory scenarios with auditory attentional modulation, and can be decoded using electroencephalographic (EEG) data. Speech segmentation based on the relative root-mean-square (RMS) intensity can be used to estimate segmental contributions to perception in noisy conditions. High-RMS-level segments contain crucial information for speech perception. Hence, this study aimed to investigate the effect of high-RMS-level speech segments on auditory attention decoding performance under various signal-to-noise ratio (SNR) conditions. Scalp EEG signals were recorded when subjects listened to the attended speech stream in the mixed speech narrated concurrently by two Mandarin speakers. The temporal response function was used to identify the attended speech from EEG responses of tracking to the temporal envelopes of intact speech and high-RMS-level speech segments alone, respectively. Auditory decoding performance was then analyzed under various SNR conditions by comparing EEG correlations to the attended and ignored speech streams. The accuracy of auditory attention decoding based on the temporal envelope with high-RMS-level speech segments was not inferior to that based on the temporal envelope of intact speech. Cortical activity correlated more strongly with attended than with ignored speech under different SNR conditions. These results suggest that EEG recordings corresponding to high-RMS-level speech segments carry crucial information for the identification and tracking of attended speech in the presence of background noise. This study also showed that with the modulation of auditory attention, attended speech can be decoded more robustly from neural activity than from behavioral measures under a wide range of SNR.
Keywords: EEG; auditory attention decoding; signal-to-noise ratio; speech RMS-level segments; temporal response function (TRF).
Copyright © 2020 Wang, Wu and Chen.
Figures






Similar articles
-
A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021. Front Neurosci. 2022. PMID: 35221885 Free PMC article.
-
EEG-based auditory attention decoding using speech-level-based segmented computational models.J Neural Eng. 2021 May 25;18(4). doi: 10.1088/1741-2552/abfeba. J Neural Eng. 2021. PMID: 33957606
-
Cortical auditory responses index the contributions of different RMS-level-dependent segments to speech intelligibility.Hear Res. 2019 Nov;383:107808. doi: 10.1016/j.heares.2019.107808. Epub 2019 Oct 4. Hear Res. 2019. PMID: 31606583
-
Noise-robust cortical tracking of attended speech in real-world acoustic scenes.Neuroimage. 2017 Aug 1;156:435-444. doi: 10.1016/j.neuroimage.2017.04.026. Epub 2017 Apr 13. Neuroimage. 2017. PMID: 28412441
-
Neural Encoding of Attended Continuous Speech under Different Types of Interference.J Cogn Neurosci. 2018 Nov;30(11):1606-1619. doi: 10.1162/jocn_a_01303. Epub 2018 Jul 13. J Cogn Neurosci. 2018. PMID: 30004849 Review.
Cited by
-
Auditory Attention Detection via Cross-Modal Attention.Front Neurosci. 2021 Jul 21;15:652058. doi: 10.3389/fnins.2021.652058. eCollection 2021. Front Neurosci. 2021. PMID: 34366770 Free PMC article.
-
A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.Front Neurosci. 2022 Feb 10;15:760611. doi: 10.3389/fnins.2021.760611. eCollection 2021. Front Neurosci. 2022. PMID: 35221885 Free PMC article.
-
A Brain-Computer Interface for Improving Auditory Attention in Multi-Talker Environments.bioRxiv [Preprint]. 2025 Mar 13:2025.03.13.641661. doi: 10.1101/2025.03.13.641661. bioRxiv. 2025. PMID: 40161643 Free PMC article. Preprint.
References
LinkOut - more resources
Full Text Sources