Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Jan-Dec:28:23312165241287622.
doi: 10.1177/23312165241287622.

Processing of Visual Speech Cues in Speech-in-Noise Comprehension Depends on Working Memory Capacity and Enhances Neural Speech Tracking in Older Adults With Hearing Impairment

Affiliations

Processing of Visual Speech Cues in Speech-in-Noise Comprehension Depends on Working Memory Capacity and Enhances Neural Speech Tracking in Older Adults With Hearing Impairment

Vanessa Frei et al. Trends Hear. 2024 Jan-Dec.

Abstract

Comprehending speech in noise (SiN) poses a challenge for older hearing-impaired listeners, requiring auditory and working memory resources. Visual speech cues provide additional sensory information supporting speech understanding, while the extent of such visual benefit is characterized by large variability, which might be accounted for by individual differences in working memory capacity (WMC). In the current study, we investigated behavioral and neurofunctional (i.e., neural speech tracking) correlates of auditory and audio-visual speech comprehension in babble noise and the associations with WMC. Healthy older adults with hearing impairment quantified by pure-tone hearing loss (threshold average: 31.85-57 dB, N = 67) listened to sentences in babble noise in audio-only, visual-only and audio-visual speech modality and performed a pattern matching and a comprehension task, while electroencephalography (EEG) was recorded. Behaviorally, no significant difference in task performance was observed across modalities. However, we did find a significant association between individual working memory capacity and task performance, suggesting a more complex interplay between audio-visual speech cues, working memory capacity and real-world listening tasks. Furthermore, we found that the visual speech presentation was accompanied by increased cortical tracking of the speech envelope, particularly in a right-hemispheric auditory topographical cluster. Post-hoc, we investigated the potential relationships between the behavioral performance and neural speech tracking but were not able to establish a significant association. Overall, our results show an increase in neurofunctional correlates of speech associated with congruent visual speech cues, specifically in a right auditory cluster, suggesting multisensory integration.

Keywords: EEG; age-related hearing loss; audio-visual speech; neural speech tracking; speech in noise; working memory capacity.

PubMed Disclaimer

Conflict of interest statement

Declaration of Conflicting InterestsThe authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Figures

Figure 1.
Figure 1.
Pure-tone audiometry. The audiogram depicts individual pure-tone thresholds at frequencies between 0.5 and 8 kHz. There is no systematic difference between hearing-aid users (HA; group average depicted in red) and non-hearing-aid users (nHA; group average depicted in blue). Stimulus presentation was limited to 100 dB, which explains the accumulation of data points at 8 kHz. Hearing-aid users were measured while having their devices on.
Figure 2.
Figure 2.
Illustration of the stimulus presentation. The modalities differed in that audio-visual-babble (AVB) included a video sequence of the mouth and jaw movement whereas the audio-babble (AB) modality only contained auditory stimuli. Five sentences were presented, and after each, a pattern-matching task was applied. After every fifth sentence, a comprehension question was asked. There was a total of 30 items per modality.
Figure 3.
Figure 3.
Speech in noise performance estimated by speech presentation modality, working memory capacity, age, speech tracking and hearing loss. A: No significant increase in pattern matching across the two speech presentation modalities was revealed by the model. B: No significant increase in comprehension performance across modalities was observed. Compared to pattern matching, for the comprehension task only after every 5th trial, a comprehension question was asked, resulting in a six-level classification of performance (30 trials within each modality were presented in total). C: Working memory capacity was significantly positively associated with pattern matching, while age and hearing loss revealed the opposite relationship. Neural speech tracking was not significantly associated. D: Working memory capacity was significantly associated with comprehension performance, while neither age, neural speech tracking nor hearing loss explained variance in the comprehension performance. n.s. = not significant, *p < .05, **p < .01, ***p < .001.
Figure 4.
Figure 4.
Topographic distribution and time course of neural speech tracking. A: Grand average cross-correlation functions of the right, the left and the frontal cluster. Significant time windows are marked as bars over the function. B: Topographic distribution and time course of the grand average cross-correlation in all three listening modalities from approximately 50 to 250 ms. C: Topographical distribution of the grand-average cross-correlation at the peaks at approximately 100 and 250 ms. Selected electrode clusters are marked with “•”. Warm colors denote positive- and cool colors negative correlations.
Figure 5.
Figure 5.
Neural speech tracking across modalities and cluster. Neural speech tracking significantly differed across modalities. In the presence of congruent visual speech cues, cross-correlation coefficients increased, particularly in the right auditory-related cluster. n.s. = not significant, *p < .05, **p < .01, ***p < .001.

Similar articles

References

    1. Abrams D. A., Nicol T., Zecker S., Kraus N. (2008). Right-Hemisphere auditory cortex is dominant for coding syllable patterns in speech. Journal of Neuroscience, 28(15), 3958–3965. 10.1523/JNEUROSCI.0187-08.2008 - DOI - PMC - PubMed
    1. Akeroyd M. A. (2008). Are individual differences in speech reception related to individual differences in cognitive ability? A survey of twenty experimental studies with normal and hearing-impaired adults. International Journal of Audiology, 47(sup2), S53–S71. 10.1080/14992020802301142 - DOI - PubMed
    1. Aller M., Økland H. S., MacGregor L. J., Blank H., Davis M. H. (2022). Differential auditory and visual phase-locking are observed during audio-visual benefit and silent lip-Reading for speech perception. Journal of Neuroscience, 42(31), 6108–6120. 10.1523/JNEUROSCI.2476-21.2022 - DOI - PMC - PubMed
    1. Altieri N., Hudock D. (2014). Hearing impairment and audiovisual speech integration ability: A case study report. Frontiers in Psychology, 5, 1–11. 10.3389/fpsyg.2014.00678 . - DOI - PMC - PubMed
    1. Anderson S., Parbery-Clark A., Yi H.-G., Kraus N. (2011). A neural basis of speech-in-noise perception in older adults. Ear and Hearing, 32(6), 750–757, 10.1097/AUD.0b013e31822229d3 - DOI - PMC - PubMed

Publication types

LinkOut - more resources