A model for context effects in speech recognition
- PMID: 8423265
- DOI: 10.1121/1.406844
A model for context effects in speech recognition
Abstract
A model is presented that quantifies the effect of context on speech recognition. In this model, a speech stimulus is considered as a concatenation of a number of equivalent elements (e.g., phonemes constituting a word). The model employs probabilities that individual elements are recognized and chances that missed elements are guessed using contextual information. Predictions are given of the probability that the entire stimulus, or part of it, is reproduced correctly. The model can be applied to both speech recognition and visual recognition of printed text. It has been verified with data obtained with syllables of the consonant-vowel-consonant (CVC) type presented near the reception threshold in quiet and in noise, with the results of an experiment using orthographic presentation of incomplete CVC syllables and with results of word counts in a CVC lexicon. A remarkable outcome of the analysis is that the cues which occur only in spoken language (e.g., coarticulatory cues) seem to have a much greater influence on recognition performance when the stimuli are presented near the threshold in noise than when they are presented near the absolute threshold. Demonstrations are given of further predictions provided by the model: word recognition as a function of signal-to-noise ratio, closed-set word recognition, recognition of interrupted speech, and sentence recognition.
Similar articles
-
Speech identification in noise: Contribution of temporal, spectral, and visual speech cues.J Acoust Soc Am. 2009 Dec;126(6):3246-57. doi: 10.1121/1.3250425. J Acoust Soc Am. 2009. PMID: 20000938
-
Congruent and incongruent semantic context influence vowel recognition.Lang Speech. 2011 Sep;54(Pt 3):341-60. doi: 10.1177/0023830911402476. Lang Speech. 2011. PMID: 22070043
-
Large-corpus phoneme and word recognition and the generality of lexical context in CVC word perception.J Speech Lang Hear Res. 2014 Feb;57(1):297-307. doi: 10.1044/1092-4388(2013/12-0183). J Speech Lang Hear Res. 2014. PMID: 24687475
-
The influence of semantically related and unrelated text cues on the intelligibility of sentences in noise.Ear Hear. 2011 Nov-Dec;32(6):e16-25. doi: 10.1097/AUD.0b013e318228036a. Ear Hear. 2011. PMID: 21826004
-
The benefit obtained from visually displayed text from an automatic speech recognizer during listening to speech presented in noise.Ear Hear. 2008 Dec;29(6):838-52. doi: 10.1097/AUD.0b013e31818005bd. Ear Hear. 2008. PMID: 18633325
Cited by
-
Evaluation of Speech-Perception Training for Hearing Aid Users: A Multisite Study in Progress.Semin Hear. 2015 Nov;36(4):273-83. doi: 10.1055/s-0035-1564453. Semin Hear. 2015. PMID: 27587914 Free PMC article.
-
Efficiency in glimpsing vowel sequences in fluctuating makers: Effects of temporal fine structure and temporal regularity.J Acoust Soc Am. 2019 Apr;145(4):2518. doi: 10.1121/1.5098949. J Acoust Soc Am. 2019. PMID: 31046353 Free PMC article.
-
Sentence perception in noise by hearing-aid users predicted by syllable-constituent perception and the use of context.J Acoust Soc Am. 2020 Jan;147(1):273. doi: 10.1121/10.0000563. J Acoust Soc Am. 2020. PMID: 32006979 Free PMC article.
-
Development of a Phrase-Based Speech-Recognition Test Using Synthetic Speech.Trends Hear. 2024 Jan-Dec;28:23312165241261490. doi: 10.1177/23312165241261490. Trends Hear. 2024. PMID: 39051703 Free PMC article.
-
The process of spoken word recognition in the face of signal degradation.J Exp Psychol Hum Percept Perform. 2014 Feb;40(1):308-27. doi: 10.1037/a0034353. Epub 2013 Sep 16. J Exp Psychol Hum Percept Perform. 2014. PMID: 24041330 Free PMC article. Clinical Trial.
MeSH terms
LinkOut - more resources
Full Text Sources
Other Literature Sources