Seeing to hear better: evidence for early audio-visual interactions in speech identification
- PMID: 15147940
- DOI: 10.1016/j.cognition.2004.01.006
Seeing to hear better: evidence for early audio-visual interactions in speech identification
Abstract
Lip reading is the ability to partially understand speech by looking at the speaker's lips. It improves the intelligibility of speech in noise when audio-visual perception is compared with audio-only perception. A recent set of experiments showed that seeing the speaker's lips also enhances sensitivity to acoustic information, decreasing the auditory detection threshold of speech embedded in noise [J. Acoust. Soc. Am. 109 (2001) 2272; J. Acoust. Soc. Am. 108 (2000) 1197]. However, detection is different from comprehension, and it remains to be seen whether improved sensitivity also results in an intelligibility gain in audio-visual speech perception. In this work, we use an original paradigm to show that seeing the speaker's lips enables the listener to hear better and hence to understand better. The audio-visual stimuli used here could not be differentiated by lip reading per se since they contained exactly the same lip gesture matched with different compatible speech sounds. Nevertheless, the noise-masked stimuli were more intelligible in the audio-visual condition than in the audio-only condition due to the contribution of visual information to the extraction of acoustic cues. Replacing the lip gesture by a non-speech visual input with exactly the same time course, providing the same temporal cues for extraction, removed the intelligibility benefit. This early contribution to audio-visual speech identification is discussed in relationships with recent neurophysiological data on audio-visual perception.
Similar articles
-
Effects of audio-visual integration on the detection of masked speech and non-speech sounds.Brain Cogn. 2011 Feb;75(1):60-6. doi: 10.1016/j.bandc.2010.09.005. Epub 2010 Nov 9. Brain Cogn. 2011. PMID: 21067852
-
Disruptive effects of auditory signal delay on speech perception with lipreading.J Aud Res. 1986 Jan;26(1):27-41. J Aud Res. 1986. PMID: 3610989
-
Bimodal audio-visual training enhances auditory adaptation process.Neuroreport. 2009 Sep 23;20(14):1231-4. doi: 10.1097/WNR.0b013e32832fbef8. Neuroreport. 2009. PMID: 19629016
-
Lipreading and audio-visual speech perception.Philos Trans R Soc Lond B Biol Sci. 1992 Jan 29;335(1273):71-8. doi: 10.1098/rstb.1992.0009. Philos Trans R Soc Lond B Biol Sci. 1992. PMID: 1348140 Review.
-
Effects of phonetic context on audio-visual intelligibility of French.J Speech Hear Res. 1994 Oct;37(5):1195-203. doi: 10.1044/jshr.3705.1195. J Speech Hear Res. 1994. PMID: 7823561 Review.
Cited by
-
Intelligibility of speech produced by sighted and blind adults.PLoS One. 2022 Sep 15;17(9):e0272127. doi: 10.1371/journal.pone.0272127. eCollection 2022. PLoS One. 2022. PMID: 36107945 Free PMC article.
-
Cross-modal prediction in speech depends on prior linguistic experience.Exp Brain Res. 2013 Apr;225(4):499-511. doi: 10.1007/s00221-012-3390-3. Epub 2013 Feb 6. Exp Brain Res. 2013. PMID: 23386124
-
Temporal synchrony and audiovisual integration of speech and object stimuli in autism.Res Autism Spectr Disord. 2017 Jul;39:11-19. doi: 10.1016/j.rasd.2017.04.001. Epub 2017 Apr 15. Res Autism Spectr Disord. 2017. PMID: 30220908 Free PMC article.
-
Causal inference of asynchronous audiovisual speech.Front Psychol. 2013 Nov 13;4:798. doi: 10.3389/fpsyg.2013.00798. eCollection 2013. Front Psychol. 2013. PMID: 24294207 Free PMC article.
-
Talking points: A modulating circle reduces listening effort without improving speech recognition.Psychon Bull Rev. 2019 Feb;26(1):291-297. doi: 10.3758/s13423-018-1489-7. Psychon Bull Rev. 2019. PMID: 29790122 Clinical Trial.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources