Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2014 Apr;111(7):1400-8.
doi: 10.1152/jn.00690.2013. Epub 2014 Jan 8.

The cortical representation of the speech envelope is earlier for audiovisual speech than audio speech

Affiliations
Free article

The cortical representation of the speech envelope is earlier for audiovisual speech than audio speech

Michael J Crosse et al. J Neurophysiol. 2014 Apr.
Free article

Retraction in

  • Retraction.
    Crosse MJ, Lalor EC. Crosse MJ, et al. J Neurophysiol. 2014 Nov 15;112(10):2667. doi: 10.1152/jn.z9k-2710-retr.2014. J Neurophysiol. 2014. PMID: 25399450 No abstract available.

Abstract

Visual speech can greatly enhance a listener's comprehension of auditory speech when they are presented simultaneously. Efforts to determine the neural underpinnings of this phenomenon have been hampered by the limited temporal resolution of hemodynamic imaging and the fact that EEG and magnetoencephalographic data are usually analyzed in response to simple, discrete stimuli. Recent research has shown that neuronal activity in human auditory cortex tracks the envelope of natural speech. Here, we exploit this finding by estimating a linear forward-mapping between the speech envelope and EEG data and show that the latency at which the envelope of natural speech is represented in cortex is shortened by >10 ms when continuous audiovisual speech is presented compared with audio-only speech. In addition, we use a reverse-mapping approach to reconstruct an estimate of the speech stimulus from the EEG data and, by comparing the bimodal estimate with the sum of the unimodal estimates, find no evidence of any nonlinear additive effects in the audiovisual speech condition. These findings point to an underlying mechanism that could account for enhanced comprehension during audiovisual speech. Specifically, we hypothesize that low-level acoustic features that are temporally coherent with the preceding visual stream may be synthesized into a speech object at an earlier latency, which may provide an extended period of low-level processing before extraction of semantic information.

Keywords: EEG; TRF; analysis-by-synthesis; latency; multisensory integration.

PubMed Disclaimer

Publication types

LinkOut - more resources