Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2017 Jun 2;7(6):60.
doi: 10.3390/brainsci7060060.

Electrophysiological Indices of Audiovisual Speech Perception in the Broader Autism Phenotype

Affiliations

Electrophysiological Indices of Audiovisual Speech Perception in the Broader Autism Phenotype

Julia Irwin et al. Brain Sci. .

Abstract

When a speaker talks, the consequences of this can both be heard (audio) and seen (visual). A novel visual phonemic restoration task was used to assess behavioral discrimination and neural signatures (event-related potentials, or ERP) of audiovisual processing in typically developing children with a range of social and communicative skills assessed using the social responsiveness scale, a measure of traits associated with autism. An auditory oddball design presented two types of stimuli to the listener, a clear exemplar of an auditory consonant-vowel syllable /ba/ (the more frequently occurring standard stimulus), and a syllable in which the auditory cues for the consonant were substantially weakened, creating a stimulus which is more like /a/ (the infrequently presented deviant stimulus). All speech tokens were paired with a face producing /ba/ or a face with a pixelated mouth containing motion but no visual speech. In this paradigm, the visual /ba/ should cause the auditory /a/ to be perceived as /ba/, creating an attenuated oddball response; in contrast, a pixelated video (without articulatory information) should not have this effect. Behaviorally, participants showed visual phonemic restoration (reduced accuracy in detecting deviant /a/) in the presence of a speaking face. In addition, ERPs were observed in both an early time window (N100) and a later time window (P300) that were sensitive to speech context (/ba/ or /a/) and modulated by face context (speaking face with visible articulation or with pixelated mouth). Specifically, the oddball responses for the N100 and P300 were attenuated in the presence of a face producing /ba/ relative to a pixelated face, representing a possible neural correlate of the phonemic restoration effect. Notably, those individuals with more traits associated with autism (yet still in the non-clinical range) had smaller P300 responses overall, regardless of face context, suggesting generally reduced phonemic discrimination.

Keywords: ERP; audiovisual speech perception; broader autism phenotype; development.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure A1
Figure A1
Frequency distribution of useable trials as a function of face context and speech stimulus.
Figure 1
Figure 1
Spectrogram of /ba/ and /a/ synthesized auditory speech stimuli. First panel, top left, spectrogram of synthesized /ba/; Second panel, top right, edited synthesized /ba/ with reduced initial formants for the consonant, referred to as /a/.
Figure 2
Figure 2
Image of audiovisual (AV) and pixelated (PX) face condition stimuli. Left panel audiovisual face condition, showing the visible articulation of the speaker; Right panel pixelated face condition, showing the speaker’s face, but obscuring the mouth.
Figure 3
Figure 3
Waveform plots showing the N100 and P300. (A): electrode montage; (B) and (C): N100 and P300 response to standard /ba/ and deviant /a/ in the audiovisual (AV) condition; Right panel: N100 and P300 response to standard /ba/ and deviant /a/ in the pixelated video (PX) condition. Shading around waveforms represents the standard error from the mean.
Figure 4
Figure 4
Topomaps for the N100 and P300 effects by condition.
Figure 5
Figure 5
Scatter plots showing the correlation between the late P300 effect and the social responsiveness scale (SRS) total score in the AV and PX conditions.

References

    1. Sumby W.H., Pollack I. Visual contribution to speech intelligibility in noise. J. Acoust. Soc. Am. 1954;26:211–215. doi: 10.1121/1.1907309. - DOI
    1. McGurk H., MacDonald J. Hearing lips and seeing voices. Nature. 1976;264:741–748. doi: 10.1038/264746a0. - DOI - PubMed
    1. Burnham D., Dodd B. Familiarity and novelty preferences in infants’ auditory-visual speech perception: Problems, factors, and a solution. Adv. Inf. Res. 1998;12:171–187.
    1. Desjardins R.N., Werker J.F. Is the integration of heard and seen speech mandatory for infants? Dev. Psychobiol. 2004;45:181–203. doi: 10.1002/dev.20033. - DOI - PubMed
    1. Meltzoff A.N., Kuhl P.K. Faces and speech: Intermodal processing of biologically relevant signals in infants and adults. Dev. Intersens. Percept. 1994;19:331–369.

LinkOut - more resources