The effect of lip-reading on primary stream segregation
- PMID: 21786898
- PMCID: PMC3155588
- DOI: 10.1121/1.3592223
The effect of lip-reading on primary stream segregation
Abstract
Lip-reading has been shown to improve the intelligibility of speech in multitalker situations, where auditory stream segregation naturally takes place. This study investigated whether the benefit of lip-reading is a result of a primary audiovisual interaction that enhances the obligatory streaming mechanism. Two behavioral experiments were conducted involving sequences of French vowels that alternated in fundamental frequency. In Experiment 1, subjects attempted to identify the order of items in a sequence. In Experiment 2, subjects attempted to detect a disruption to temporal isochrony across alternate items. Both tasks are disrupted by streaming, thus providing a measure of primary or obligatory streaming. Visual lip gestures articulating alternate vowels were synchronized with the auditory sequence. Overall, the results were consistent with the hypothesis that visual lip gestures enhance segregation by affecting primary auditory streaming. Moreover, increases in the naturalness of visual lip gestures and auditory vowels, and corresponding increases in audiovisual congruence may potentially lead to increases in the effect of visual lip gestures on streaming.
© 2011 Acoustical Society of America
Figures







Similar articles
-
Temporally pre-presented lipreading cues release speech from informational masking.J Acoust Soc Am. 2013 Apr;133(4):EL281-5. doi: 10.1121/1.4794933. J Acoust Soc Am. 2013. PMID: 23556692
-
The effect of audiovisual and binaural listening on the acceptable noise level (ANL): establishing an ANL conceptual model.J Am Acad Audiol. 2014 Feb;25(2):141-53. doi: 10.3766/jaaa.25.2.3. J Am Acad Audiol. 2014. PMID: 24828215
-
Sentence intelligibility during segmental interruption and masking by speech-modulated noise: Effects of age and hearing loss.J Acoust Soc Am. 2015 Jun;137(6):3487-501. doi: 10.1121/1.4921603. J Acoust Soc Am. 2015. PMID: 26093436 Free PMC article.
-
Effects of phonetic context on audio-visual intelligibility of French.J Speech Hear Res. 1994 Oct;37(5):1195-203. doi: 10.1044/jshr.3705.1195. J Speech Hear Res. 1994. PMID: 7823561 Review.
-
Look 'hear', primary auditory cortex is active during lip-reading.Neuroreport. 2005 Feb 8;16(2):123-4. doi: 10.1097/00001756-200502080-00009. Neuroreport. 2005. PMID: 15671859 Review. No abstract available.
Cited by
-
The effect of visual cues on difficulty ratings for segregation of musical streams in listeners with impaired hearing.PLoS One. 2011;6(12):e29327. doi: 10.1371/journal.pone.0029327. Epub 2011 Dec 15. PLoS One. 2011. PMID: 22195046 Free PMC article.
-
Schlieren imaging and video classification of alphabet pronunciations: exploiting phonetic flows for speech recognition and speech therapy.Vis Comput Ind Biomed Art. 2024 May 22;7(1):12. doi: 10.1186/s42492-024-00163-w. Vis Comput Ind Biomed Art. 2024. PMID: 38772963 Free PMC article.
-
Auditory selective attention is enhanced by a task-irrelevant temporally coherent visual stimulus in human listeners.Elife. 2015 Feb 5;4:e04995. doi: 10.7554/eLife.04995. Elife. 2015. PMID: 25654748 Free PMC article.
-
The Effect of Sound Localization on Auditory-Only and Audiovisual Speech Recognition in a Simulated Multitalker Environment.Trends Hear. 2023 Jan-Dec;27:23312165231186040. doi: 10.1177/23312165231186040. Trends Hear. 2023. PMID: 37415497 Free PMC article.
-
Training enhances the ability of listeners to exploit visual information for auditory scene analysis.Cognition. 2021 Mar;208:104529. doi: 10.1016/j.cognition.2020.104529. Epub 2020 Dec 26. Cognition. 2021. PMID: 33373937 Free PMC article.
References
-
- American National Standards Institute (1995). ANSI S3.7-R2003: Methods for Coupler Calibration of Earphones, American National Standards Institute, NY.
-
- American National Standards Institute (2004). ANSI S3.21-2004: Methods for Manual Pure-Tone Threshold Audiometry, American National Standards Institute, NY.
-
- Bernstein, L. E., Auer, E. T. J., and Takayanagi, S. (2004). “Auditory speech detection in noise enhanced by lipreading,” Speech Commun. 44, 5–18. 10.1016/j.specom.2004.10.011 - DOI
-
- Berthommier, F. (2003). “A phonetically neutral model of the low-level audiovisual interaction,” in Proceedings of the International Conference on Audio-Visual Speech Processing, 89–94 (Institut de la Communication Parlée, St. Jorioz, France: ).
Publication types
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
Medical