Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2020 Jan-Dec:24:2331216520948390.
doi: 10.1177/2331216520948390.

Sound Externalization: A Review of Recent Research

Affiliations
Review

Sound Externalization: A Review of Recent Research

Virginia Best et al. Trends Hear. 2020 Jan-Dec.

Abstract

Sound externalization, or the perception that a sound source is outside of the head, is an intriguing phenomenon that has long interested psychoacousticians. While previous reviews are available, the past few decades have produced a substantial amount of new data.In this review, we aim to synthesize those data and to summarize advances in our understanding of the phenomenon. We also discuss issues related to the definition and measurement of sound externalization and describe quantitative approaches that have been taken to predict the outcomes of externalization experiments. Last, sound externalization is of practical importance for many kinds of hearing technologies. Here, we touch on two examples, discussing the role of sound externalization in augmented/virtual reality systems and bringing attention to the somewhat overlooked issue of sound externalization in wearers of hearing aids.

Keywords: distal attribution; in-head localization; sound image externalization; spatial perception; virtual acoustics.

PubMed Disclaimer

Figures

Figure 1.
Figure 1.
Sound externalization (A) as a binary decision, (B) as a continuum, and (C) depending on lateral position.
Figure 2.
Figure 2.
Mean externalization ratings from three studies showing the tendency for increased externalization with sound source laterality. Brimijoin et al. (2013, N = 11): Ratings correspond to the proportion of externalized responses obtained using a binary judgment and are averaged across fullband and lowpass conditions. Small head movements were allowed, and the sources were either fixed in azimuth (“world-fixed”) or moved along with the head movements (“head-fixed”). Kates et al. (2018, N = 20): Ratings were obtained using a 100-point scale with visual references and are shown separately for anechoic and reverberant conditions. Leclère et al. (2019, N = 21): Ratings were obtained using a continuous percentage scale with eyes closed and are shown separately for anechoic and reverberant conditions (the latter averaged across four rooms). Error bars show standard errors of the mean.
Figure 3.
Figure 3.
Effects of spectral smoothing methods on an example HRTF at 60° azimuth. Top row: monaural spectrum at the ipsilateral ear. Bottom row: difference between the left- and right-ear spectra (or the ILD). Spectral representations were obtained by filtering the impulse response with a gammatone filter bank (Lyon, 1997) and calculating the root mean square amplitude within each band. Left column: smoothing according to Hassager et al. (2016) where B is the spectral smoothing bandwidth relative to one ERBN (Glasberg & Moore, 1990). Right column: smoothing according to Baumgartner et al. (2017) where C is a scaling factor that is applied to spectral magnitudes. Gray-shaded areas indicate frequency ranges not tested in the respective experiments. The HRTF is from subject S01 in Baumgartner et al. (2017). ILD = interaural level difference.
Figure 4.
Figure 4.
General structure of template-based externalization models that calculate the deviation of acoustical cues from those of a natural, well-externalized, reference stimulus. Models focused on the direct sound use long-term excitation patterns and involve comparisons of monaural cues at each ear and/or interaural cues (e.g., Baumgartner & Majdak, 2020; Hassager et al. 2016). Models focused on reverberation use short-term excitation patterns and involve only interaural-cue comparisons (e.g., Li et al. 2018).

References

    1. Akeroyd M. A., Chambers J., Bullock D., Palmer A. R., Summerfield Q., Nelson P. A., Gatehouse S. (2007). The binaural performance of a cross-talk cancellation system with matched or mismatched setup and playback acoustics. Journal of the Acoustical Society of America, 121(2), 1056–1069. 10.1121/1.2404625 - DOI - PMC - PubMed
    1. Baumgartner R., Majdak P. (2020). Decision making in auditory externalization perception. bioRxiv. 10.1101/2020.04.30.068817 - DOI
    1. Baumgartner R., Reed D. K., Tóth B., Best V., Majdak P., Colburn H. S., Shinn-Cunningham B. (2017). Asymmetries in behavioral and neural responses to spectral cues demonstrate the generality of auditory looming bias. Proceedings of the National Academy of Sciences of the United States of America, 114(36), 9743–9748. 10.1073/pnas.1703247114 - DOI - PMC - PubMed
    1. Begault D. R., Wenzel E. M. (1993). Headphone localization of speech. Human Factors, 35(2), 361–376. 10.1177/001872089303500210 - DOI - PubMed
    1. Begault D. R., Wenzel E. M., Anderson M. R. (2001). Direct comparison of the impact of head-tracking, reverberation, and individualized head-related transfer functions on the spatial perception of a virtual speech source. Journal of the Audio Engineering Society, 49(10), 904–916. - PubMed

Publication types