Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2013 Dec 3:2:e01312.
doi: 10.7554/eLife.01312.

Decoding neural responses to temporal cues for sound localization

Affiliations

Decoding neural responses to temporal cues for sound localization

Dan F M Goodman et al. Elife. .

Abstract

The activity of sensory neural populations carries information about the environment. This may be extracted from neural activity using different strategies. In the auditory brainstem, a recent theory proposes that sound location in the horizontal plane is decoded from the relative summed activity of two populations in each hemisphere, whereas earlier theories hypothesized that the location was decoded from the identity of the most active cells. We tested the performance of various decoders of neural responses in increasingly complex acoustical situations, including spectrum variations, noise, and sound diffraction. We demonstrate that there is insufficient information in the pooled activity of each hemisphere to estimate sound direction in a reliable way consistent with behavior, whereas robust estimates can be obtained from neural activity by taking into account the heterogeneous tuning of cells. These estimates can still be obtained when only contralateral neural responses are used, consistently with unilateral lesion studies. DOI: http://dx.doi.org/10.7554/eLife.01312.001.

Keywords: audition; neural coding; sound localization.

PubMed Disclaimer

Conflict of interest statement

The authors declare that no competing interests exist.

Figures

Figure 1.
Figure 1.. Overview of model.
(A) The distribution of best delay vs best frequency for cells in the guinea pig model (left), with the physiological range of ITDs shown in gray, and a sample tuning curve (right). (B) Illustration of the model: a sound source is filtered via position-dependent HRTFs to give left and right channels. For each best frequency on each channel, the signal undergoes cochlear filtering using gammatone filters. An internal delay is added, and the two channels are combined and sent to a binaural neuron model that produces Poisson distributed spikes. (C) The response of the cells to sounds at two different ITDs (rows) for white noise (left column) and a natural sound (right column). The ITD is indicated by the black dashed line. Each cell is surrounded by a disc with a color indicating the response of that neuron (hot colors corresponding to strong responses). When two or more discs overlap, each point is colored according to the closest cell. The strongest responses lie along the line of the ITD. DOI: http://dx.doi.org/10.7554/eLife.01312.003
Figure 2.
Figure 2.. Decoders in single frequency bands.
(A) Peak and hemispheric decoders. Left: response of binaural neurons to sound at ITD = 0.2 ms (dashed line), in a narrow frequency band. The size of points is proportionate to spike count, and the crossed point corresponds to the highest spike count. Middle: the same cell responses are displayed as best delay vs spike count (note the different horizontal axis). The solid black line is the Gaussian smoothed spike count, whose peak (circle) is the ITD estimate. The maximally responsive neuron is also indicated with a circle for comparison. The yellow and orange bars give the mean response of neurons with positive and negative best delays, respectively, from which the normalized hemispheric difference is computed. Right: the hemispheric difference as a function of ITD at 700 Hz (blue) and 1.3 kHz (purple). At 1.3 kHz, the difference shown by the dashed line gives an ambiguous estimate of the ITD. (B) Mean error for the guinea pig and cat, for the peak (blue, dashed), smoothed peak (blue, solid), hemispheric (red), and pattern match (green) decoders. The distribution of BD vs BF is shown in the inset. (C) Illustration of the pattern match decoder and a neural circuit that implements it. The response (left) is compared to two patterns A and B, corresponding to two different ITDs (right). Each response neuron is connected to a pattern-sensitive neuron with weights proportional to the stored response of each pattern. When the weights match the responses, the output of the pattern-sensitive neuron is strongest. DOI: http://dx.doi.org/10.7554/eLife.01312.004
Figure 3.
Figure 3.. Integration across frequencies.
(A) Mean error in estimating ITD for white noise using the smoothed peak (blue), hemispheric (red), and pattern match (green) decoders, as a function of the number of binaural cells. Training and testing of the decoders are both performed using white noise. (B) Mean error as a function of frequency band when decoders are trained on white noise but tested on band-pass noise centered at the given frequency. Notice the different vertical scale between (A and B). (C) Performance when cells with a frequency above the cutoff are discarded. (D) Mean error and bias to center in the decoders for guinea pig (with a maximum frequency of 1.2 kHz) when trained on white noise and tested on colored noise. DOI: http://dx.doi.org/10.7554/eLife.01312.005
Figure 3—figure supplement 1.
Figure 3—figure supplement 1.. Comparison with standard machine learning decoders.
As for Figure 3, except that the main decoders are shown in light colors, and there are two additional decoders: ridge regression (magenta) and nearest neighbor regression (black). DOI: http://dx.doi.org/10.7554/eLife.01312.006
Figure 4.
Figure 4.. Frequency-dependent improvements.
(A) Comparing hemispheric differences across frequency channels. In each plot, color indicates frequency with red being high frequency and blue being low frequency. Left: tuning curves for a few binaural neurons. Middle: hemispheric difference (L − R)/(L + R). Right: frequency-dependent hemispheric difference (1/f) (L − R)/(L + R). (B) Mean error as a function of frequency band in the guinea pig model, for hemispheric (red) and pattern match (green) decoders (dashed lines), and frequency-dependent hemispheric and pattern match (solid) decoders. The shaded regions show the difference between the simple and frequency-dependent versions. The dotted lines show the mean error for band-pass noise if the decoder is trained and tested on the same frequency band, as shown in Figure 2B (guinea pig). This represents the lower bound for the error. DOI: http://dx.doi.org/10.7554/eLife.01312.007
Figure 5.
Figure 5.. Background acoustic noise.
(A) Illustration of protocol: a binaural sound is presented with a given ITD, with independent white noise added to the two channels. (B) Mean error (left column) and central bias (right column) at different signal to noise levels. Decoders are smoothed peak (blue), hemispheric (red), and pattern match (green). DOI: http://dx.doi.org/10.7554/eLife.01312.008
Figure 6.
Figure 6.. Realistic head-related transfer functions.
(A) Photograph of stuffed guinea pig used for HRTF recordings, and three pairs of left/right ear impulse responses corresponding to the directions marked on the photograph. (B) Frequency dependence of ITD for the three azimuths shown in panel (A), in guinea pig and cat HRTFs. (C) Mean response of the model to white noise stimuli at the same azimuth (90°) for both animals, the frequency-dependent ITD curve is shown for this azimuth (dashed). (D) Performance of the model as a function of the number of cells for hemispheric (red) and pattern match (green) decoders. DOI: http://dx.doi.org/10.7554/eLife.01312.009
Figure 7.
Figure 7.. Effect of heterogeneity and lesions.
(A) Mean error for the hemispheric (red) and pattern match (green) decoders in the guinea pig model, depending on the spread of the best delays, for white noise presented with acoustic noise (SNR between −5 and 5 dB, no HRTF filtering). For every frequency, the standard deviation of BDs is multiplied by the spread factor: lower than 1 denotes less heterogeneous than the original distribution (top left), greater than 1 denotes more heterogeneous (top right). Dashed lines represent the estimation error for the original distribution. (B) Mean error for the pattern match (green) and smooth peak (blue) decoders before (dashed) and after (solid) lesioning one hemisphere in the guinea pig, as a function of presented ITD. The model is retrained after lesioning. The error curves are Gaussian smoothed to reduce noise and improve readability. DOI: http://dx.doi.org/10.7554/eLife.01312.010
Figure 8.
Figure 8.. Humans and owls.
(A) Mean error for the pattern match (green) and hemispheric (red) decoders for the barn owl model, with sounds presented through measured HRTFs. (B) Performance in the human model with uniformly distributed best interaural phase differences. (C) Performance in the human model with best delays distributed as in the guinea pig model. (D) Performance in the human model with best delays distributed as in the cat model. DOI: http://dx.doi.org/10.7554/eLife.01312.011

Similar articles

Cited by

References

    1. Aitkin LM, Pettigrew JD, Calford MB, Phillips SC, Wise LZ. 1985. Representation of stimulus azimuth by low-frequency neurons in inferior colliculus of the cat. J Neurophysiol 53:43–59 - PubMed
    1. Ashida G, Carr CE. 2011. Sound localization: jeffress and beyond. Curr Opin Neurobiol 21:745–51.10.1016/j.conb.2011.05.008 - DOI - PMC - PubMed
    1. Attwell D, Laughlin SB. 2001. An energy budget for signaling in the grey matter of the brain. J Cereb Blood Flow Metab 21:1133–45.10.1097/00004647-200110000-00001 - DOI - PubMed
    1. Brette R. 2010. On the interpretation of sensitivity analyses of neural responses. J Acoust Soc Am 128:2965–72.10.1121/1.3488311 - DOI - PubMed
    1. Briley PM, Kitterick PT, Summerfield AQ. 2013. Evidence for opponent process analysis of sound source location in humans. J Assoc Res Otolaryngol 14:83–101.10.1007/s10162-012-0356-x - DOI - PMC - PubMed

Publication types