Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2018;42(1):53-80.
doi: 10.1007/s10919-017-0267-y. Epub 2017 Sep 30.

Human Non-linguistic Vocal Repertoire: Call Types and Their Meaning

Affiliations

Human Non-linguistic Vocal Repertoire: Call Types and Their Meaning

Andrey Anikin et al. J Nonverbal Behav. 2018.

Abstract

Recent research on human nonverbal vocalizations has led to considerable progress in our understanding of vocal communication of emotion. However, in contrast to studies of animal vocalizations, this research has focused mainly on the emotional interpretation of such signals. The repertoire of human nonverbal vocalizations as acoustic types, and the mapping between acoustic and emotional categories, thus remain underexplored. In a cross-linguistic naming task (Experiment 1), verbal categorization of 132 authentic (non-acted) human vocalizations by English-, Swedish- and Russian-speaking participants revealed the same major acoustic types: laugh, cry, scream, moan, and possibly roar and sigh. The association between call type and perceived emotion was systematic but non-redundant: listeners associated every call type with a limited, but in some cases relatively wide, range of emotions. The speed and consistency of naming the call type predicted the speed and consistency of inferring the caller's emotion, suggesting that acoustic and emotional categorizations are closely related. However, participants preferred to name the call type before naming the emotion. Furthermore, nonverbal categorization of the same stimuli in a triad classification task (Experiment 2) was more compatible with classification by call type than by emotion, indicating the former's greater perceptual salience. These results suggest that acoustic categorization may precede attribution of emotion, highlighting the need to distinguish between the overt form of nonverbal signals and their interpretation by the perceiver. Both within- and between-call acoustic variation can then be modeled explicitly, bringing research on human nonverbal vocalizations more in line with the work on animal communication.

Keywords: Cross-linguistic naming study; Emotion; Non-linguistic vocalizations; Semantic spaces; Triad classification task.

PubMed Disclaimer

Figures

Fig. 1
Fig. 1
Top panel: semantic space representing naming distinctions in English, Swedish, and Russian. Text labels are positioned in prototypicality-adjusted cluster centroids. Bottom panel: cladograms of the major call types. Affinity propagation clustering with q selected manually (0.5 for English, 0.3 for Swedish, 0.45 for Russian). DK don’t know
Fig. 2
Fig. 2
Acoustic models for classifying vocalizations based on sound names chosen by English, Swedish, and Russian participants. Shaded areas show the acoustic class predicted by a multinomial regression model using two first principal components of 12 acoustic features (see Table 1). Small labels show the position of individual stimuli and their call type
Fig. 3
Fig. 3
Co-occurrence of sound names (vertical) and emotion names (horizontal) in English, Swedish, and Russian. Each cell shows the number of times a participant chose a particular sound name and emotion name to describe the same sound. The labels are ordered by hierarchical clustering, so that similar terms are placed close together, resulting in the dendrograms shown next to each table
Fig. 4
Fig. 4
The probability of expressing different levels of certainty in the chosen sound names and emotion names for all three language groups combined. Median of the posterior distribution and 95% CI
Fig. 5
Fig. 5
Model fit as a function of its dimensionality for the triad classification task. Shown: Pearson’s correlation with distance matrices from Experiment 1 based on naming the sound or emotion and normalized negative WAIC (larger is better)
Fig. 6
Fig. 6
Three-dimensional configuration of stimuli in the triad task. Each sound is labeled and color-coded by its most common name in Experiment 1 (cf. Fig. 1). Prototypicality-adjusted centroids are shown in large bold font. Origin lies in the center of gravity of each cloud. DK don’t know

References

    1. Ackermann H, Hage SR, Ziegler W. Brain mechanisms of acoustic communication in humans and nonhuman primates: An evolutionary perspective. Behavioral and Brain Sciences. 2014;37(06):529–546. doi: 10.1017/S0140525X13003099. - DOI - PubMed
    1. Alvarado N. Congruence of meaning between facial expressions of emotion and selected emotion terms. Motivation and Emotion. 1996;20(1):33–61. doi: 10.1007/BF02251006. - DOI
    1. Ameel E, Storms G, Malt BC, Sloman SA. How bilinguals solve the naming problem. Journal of Memory and Language. 2005;53(1):60–80. doi: 10.1016/j.jml.2005.02.004. - DOI
    1. Anikin A, Lima CF. Perceptual and acoustic differences between authentic and acted nonverbal emotional vocalizations. The Quarterly Journal of Experimental Psychology. 2017 - PubMed
    1. Anikin A, Persson T. Non-linguistic vocalizations from online amateur videos for emotion research: A validated corpus. Behavior Research Methods. 2017;49(2):758–771. doi: 10.3758/s13428-016-0736-y. - DOI - PubMed

LinkOut - more resources