Noise-robust cortical tracking of attended speech in real-world acoustic scenes
- PMID: 28412441
- DOI: 10.1016/j.neuroimage.2017.04.026
Noise-robust cortical tracking of attended speech in real-world acoustic scenes
Abstract
Selectively attending to one speaker in a multi-speaker scenario is thought to synchronize low-frequency cortical activity to the attended speech signal. In recent studies, reconstruction of speech from single-trial electroencephalogram (EEG) data has been used to decode which talker a listener is attending to in a two-talker situation. It is currently unclear how this generalizes to more complex sound environments. Behaviorally, speech perception is robust to the acoustic distortions that listeners typically encounter in everyday life, but it is unknown whether this is mirrored by a noise-robust neural tracking of attended speech. Here we used advanced acoustic simulations to recreate real-world acoustic scenes in the laboratory. In virtual acoustic realities with varying amounts of reverberation and number of interfering talkers, listeners selectively attended to the speech stream of a particular talker. Across the different listening environments, we found that the attended talker could be accurately decoded from single-trial EEG data irrespective of the different distortions in the acoustic input. For highly reverberant environments, speech envelopes reconstructed from neural responses to the distorted stimuli resembled the original clean signal more than the distorted input. With reverberant speech, we observed a late cortical response to the attended speech stream that encoded temporal modulations in the speech signal without its reverberant distortion. Single-trial attention decoding accuracies based on 40-50s long blocks of data from 64 scalp electrodes were equally high (80-90% correct) in all considered listening environments and remained statistically significant using down to 10 scalp electrodes and short (<30-s) unaveraged EEG segments. In contrast to the robust decoding of the attended talker we found that decoding of the unattended talker deteriorated with the acoustic distortions. These results suggest that cortical activity tracks an attended speech signal in a way that is invariant to acoustic distortions encountered in real-life sound environments. Noise-robust attention decoding additionally suggests a potential utility of stimulus reconstruction techniques in attention-controlled brain-computer interfaces.
Keywords: Acoustic simulations; Auditory attention; Cortical entrainment; Decoding; Delta rhythms; EEG; Speech; Theta rhythms.
Copyright © 2017 Elsevier Inc. All rights reserved.
Similar articles
-
Effects of Sensorineural Hearing Loss on Cortical Synchronization to Competing Speech during Selective Attention.J Neurosci. 2020 Mar 18;40(12):2562-2572. doi: 10.1523/JNEUROSCI.1936-19.2020. Epub 2020 Feb 24. J Neurosci. 2020. PMID: 32094201 Free PMC article.
-
EEG decoding of the target speaker in a cocktail party scenario: considerations regarding dynamic switching of talker location.J Neural Eng. 2019 Jun;16(3):036017. doi: 10.1088/1741-2552/ab0cf1. Epub 2019 Mar 5. J Neural Eng. 2019. PMID: 30836345
-
Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using EEG.Neuroimage. 2020 Jan 15;205:116283. doi: 10.1016/j.neuroimage.2019.116283. Epub 2019 Oct 17. Neuroimage. 2020. PMID: 31629828
-
Neural Encoding of Attended Continuous Speech under Different Types of Interference.J Cogn Neurosci. 2018 Nov;30(11):1606-1619. doi: 10.1162/jocn_a_01303. Epub 2018 Jul 13. J Cogn Neurosci. 2018. PMID: 30004849 Review.
-
The encoding of auditory objects in auditory cortex: insights from magnetoencephalography.Int J Psychophysiol. 2015 Feb;95(2):184-90. doi: 10.1016/j.ijpsycho.2014.05.005. Epub 2014 May 16. Int J Psychophysiol. 2015. PMID: 24841996 Free PMC article. Review.
Cited by
-
Convolutional neural networks can identify brain interactions involved in decoding spatial auditory attention.PLoS Comput Biol. 2024 Aug 8;20(8):e1012376. doi: 10.1371/journal.pcbi.1012376. eCollection 2024 Aug. PLoS Comput Biol. 2024. PMID: 39116183 Free PMC article.
-
Neural Representation Enhanced for Speech and Reduced for Background Noise With a Hearing Aid Noise Reduction Scheme During a Selective Attention Task.Front Neurosci. 2020 Sep 10;14:846. doi: 10.3389/fnins.2020.00846. eCollection 2020. Front Neurosci. 2020. PMID: 33071722 Free PMC article.
-
Benefits of triple acoustic beamforming during speech-on-speech masking and sound localization for bilateral cochlear-implant users.J Acoust Soc Am. 2021 May;149(5):3052. doi: 10.1121/10.0003933. J Acoust Soc Am. 2021. PMID: 34241104 Free PMC article.
-
Neural Decoding of the Speech Envelope: Effects of Intelligibility and Spectral Degradation.Trends Hear. 2024 Jan-Dec;28:23312165241266316. doi: 10.1177/23312165241266316. Trends Hear. 2024. PMID: 39183533 Free PMC article.
-
Age-Related Differences in EEG-Based Speech Reception Threshold Estimation Using Scalp and Ear-EEG.Trends Hear. 2025 Jan-Dec;29:23312165251372462. doi: 10.1177/23312165251372462. Epub 2025 Aug 25. Trends Hear. 2025. PMID: 40853325 Free PMC article.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Other Literature Sources