Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2015 Sep 1;10(9):e0136675.
doi: 10.1371/journal.pone.0136675. eCollection 2015.

Path Models of Vocal Emotion Communication

Affiliations

Path Models of Vocal Emotion Communication

Tanja Bänziger et al. PLoS One. .

Abstract

We propose to use a comprehensive path model of vocal emotion communication, encompassing encoding, transmission, and decoding processes, to empirically model data sets on emotion expression and recognition. The utility of the approach is demonstrated for two data sets from two different cultures and languages, based on corpora of vocal emotion enactment by professional actors and emotion inference by naïve listeners. Lens model equations, hierarchical regression, and multivariate path analysis are used to compare the relative contributions of objectively measured acoustic cues in the enacted expressions and subjective voice cues as perceived by listeners to the variance in emotion inference from vocal expressions for four emotion families (fear, anger, happiness, and sadness). While the results confirm the central role of arousal in vocal emotion communication, the utility of applying an extended path modeling framework is demonstrated by the identification of unique combinations of distal cues and proximal percepts carrying information about specific emotion families, independent of arousal. The statistical models generated show that more sophisticated acoustic parameters need to be developed to explain the distal underpinnings of subjective voice quality percepts that account for much of the variance in emotion inference, in particular voice instability and roughness. The general approach advocated here, as well as the specific results, open up new research strategies for work in psychology (specifically emotion and social perception research) and engineering and computer science (specifically research and development in the domain of affective computing, particularly on automatic emotion detection and synthetic emotion expression in avatars).

PubMed Disclaimer

Conflict of interest statement

Competing Interests: The authors have declared that no competing interests exist.

Figures

Fig 1
Fig 1. The tripartite emotion expression and perception (TEEP) model (based on Brunswik's lens model).
The terms “push” and “pull” refer to the internal and the external determinants of the emotional expression, respectively, distinguished in the lower and upper parts of the figure. D = distal cues; P = percepts. Adapted from p. 120 in Scherer [36].
Fig 2
Fig 2. Graphic illustration for the Lens Model Equation.
Fig 3
Fig 3. Graphic illustration for an extended model (path analysis with separate distal and proximal cues).
Fig 4
Fig 4. Conceptual representation of the TEEP path model.
Fig 5
Fig 5. Standardized path coefficients of the estimated model for anger (data merged for MUC and GVA).
Only significant path coefficients are shown (p < .02). Significant paths with an absolute value > .2 are depicted in black, significant paths with an absolute value < .2 are depicted in black. int. = intensity; r. energy = relative energy.
Fig 6
Fig 6. Standardized path coefficients of the estimated model for arousal (data merged for MUC and GVA).
Only significant path coefficients are shown (p < .02). Significant paths with an absolute value > .2 are depicted in black. Significant paths with an absolute value < .2 are depicted in black. int. = intensity; r. energy = relative energy.
Fig 7
Fig 7. Standardized path coefficients of the estimated model for fear (data merged for MUC and GVA).
Only significant path coefficients are shown (p < .02). Significant paths with an absolute value > .2 are depicted in black. Significant paths with an absolute value < .2 are depicted in black.
Fig 8
Fig 8. Standardized path coefficients of the estimated model for sadness (data merged for MUC and GVA).
Only significant path coefficients are shown (p < .02). Significant paths with an absolute value > .2 are depicted in black. Significant paths with an absolute value < .2 are depicted in black.

References

    1. Scherer KR. Vocal affect expression: A review and a model for future research. Psych. Bull. 1986; 99: 143–165. - PubMed
    1. Scherer KR. The dynamic architecture of emotion: evidence for the component process model. Cogn Emot. 2009;23: 1307–1351.
    1. Juslin PN, Laukka P. Communication of emotions in vocal expression and music performance: different channels, same code? Psychol Bull. 2003;129: 770–814. - PubMed
    1. Scherer KR. Vocal communication of emotion: a review of research paradigms. Speech Commun. 2003;40: 227–256.
    1. Scherer KR, Clark-Polner E, Mortillaro M. In the eye of the beholder? Universality and cultural specificity in the expression and perception of emotion. Int J Psychol. 2011;46: 401–435. 10.1080/00207594.2011.626049 - DOI - PubMed

Publication types

LinkOut - more resources