Attention Drives Visual Processing and Audiovisual Integration During Multimodal Communication
- PMID: 38199864
- PMCID: PMC10919203
- DOI: 10.1523/JNEUROSCI.0870-23.2023
Attention Drives Visual Processing and Audiovisual Integration During Multimodal Communication
Abstract
During communication in real-life settings, our brain often needs to integrate auditory and visual information and at the same time actively focus on the relevant sources of information, while ignoring interference from irrelevant events. The interaction between integration and attention processes remains poorly understood. Here, we use rapid invisible frequency tagging and magnetoencephalography to investigate how attention affects auditory and visual information processing and integration, during multimodal communication. We presented human participants (male and female) with videos of an actress uttering action verbs (auditory; tagged at 58 Hz) accompanied by two movie clips of hand gestures on both sides of fixation (attended stimulus tagged at 65 Hz; unattended stimulus tagged at 63 Hz). Integration difficulty was manipulated by a lower-order auditory factor (clear/degraded speech) and a higher-order visual semantic factor (matching/mismatching gesture). We observed an enhanced neural response to the attended visual information during degraded speech compared to clear speech. For the unattended information, the neural response to mismatching gestures was enhanced compared to matching gestures. Furthermore, signal power at the intermodulation frequencies of the frequency tags, indexing nonlinear signal interactions, was enhanced in the left frontotemporal and frontal regions. Focusing on the left inferior frontal gyrus, this enhancement was specific for the attended information, for those trials that benefitted from integration with a matching gesture. Together, our results suggest that attention modulates audiovisual processing and interaction, depending on the congruence and quality of the sensory input.
Keywords: attention; audiovisual integration; magnetoencephalography (MEG); multimodal communication; neural processing; rapid invisible frequency tagging (RIFT).
Copyright © 2024 the authors.
Conflict of interest statement
The authors declare no competing financial interests.
Figures






Similar articles
-
Rapid invisible frequency tagging reveals nonlinear integration of auditory and visual information.Hum Brain Mapp. 2021 Mar;42(4):1138-1152. doi: 10.1002/hbm.25282. Epub 2020 Nov 18. Hum Brain Mapp. 2021. PMID: 33206441 Free PMC article.
-
Alpha and Beta Oscillations Index Semantic Congruency between Speech and Gestures in Clear and Degraded Speech.J Cogn Neurosci. 2018 Aug;30(8):1086-1097. doi: 10.1162/jocn_a_01301. Epub 2018 Jun 19. J Cogn Neurosci. 2018. PMID: 29916792
-
Hearing and seeing meaning in noise: Alpha, beta, and gamma oscillations predict gestural enhancement of degraded speech comprehension.Hum Brain Mapp. 2018 May;39(5):2075-2087. doi: 10.1002/hbm.23987. Epub 2018 Jan 30. Hum Brain Mapp. 2018. PMID: 29380945 Free PMC article.
-
Timing in audiovisual speech perception: A mini review and new psychophysical data.Atten Percept Psychophys. 2016 Feb;78(2):583-601. doi: 10.3758/s13414-015-1026-y. Atten Percept Psychophys. 2016. PMID: 26669309 Free PMC article. Review.
-
Some behavioral and neurobiological constraints on theories of audiovisual speech integration: a review and suggestions for new directions.Seeing Perceiving. 2011;24(6):513-39. doi: 10.1163/187847611X595864. Epub 2011 Sep 29. Seeing Perceiving. 2011. PMID: 21968081 Free PMC article. Review.
Cited by
-
Evidence for a Causal Dissociation of the McGurk Effect and Congruent Audiovisual Speech Perception via TMS.bioRxiv [Preprint]. 2023 Nov 27:2023.11.27.568892. doi: 10.1101/2023.11.27.568892. bioRxiv. 2023. Update in: Multisens Res. 2024 Aug 16;37(4-5):341-363. doi: 10.1163/22134808-bja10129. PMID: 38077093 Free PMC article. Updated. Preprint.
-
Evidence for a Causal Dissociation of the McGurk Effect and Congruent Audiovisual Speech Perception via TMS to the Left pSTS.Multisens Res. 2024 Aug 16;37(4-5):341-363. doi: 10.1163/22134808-bja10129. Multisens Res. 2024. PMID: 39191410 Free PMC article.
-
An fMRI study of crossmodal emotional congruency and the role of semantic content in the aesthetic appreciation of naturalistic art.Front Neurosci. 2025 Jul 30;19:1516070. doi: 10.3389/fnins.2025.1516070. eCollection 2025. Front Neurosci. 2025. PMID: 40809398 Free PMC article.
-
Perceptual foundation and extension to phase tagging for rapid invisible frequency tagging (RIFT).Imaging Neurosci (Camb). 2024 Jul 26;2:imag-2-00242. doi: 10.1162/imag_a_00242. eCollection 2024. Imaging Neurosci (Camb). 2024. PMID: 40800424 Free PMC article.
-
Using Rapid Invisible Frequency Tagging (RIFT) to Probe the Neural Interaction Between Representations of Speech Planning and Comprehension.Neurobiol Lang (Camb). 2025 Jul 15;6:nol_a_00171. doi: 10.1162/nol_a_00171. eCollection 2025. Neurobiol Lang (Camb). 2025. PMID: 40735639 Free PMC article.
References
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources