The auditory gating paradigm was adopted to study how much acoustic information is needed to recognize emotions from speech prosody and music performances. In Study 1, brief utterances conveying ten emotions were segmented into temporally fine-grained gates and presented to listeners, whereas Study 2 instead used musically expressed emotions. Emotion recognition accuracy increased with increasing gate duration and generally stabilized after a certain duration, with different trajectories for different emotions. Above-chance accuracy was observed for ≤100 ms stimuli for anger, happiness, neutral, and sadness, and for ≤250 ms stimuli for most other emotions, for both speech and music. This suggests that emotion recognition is a fast process that allows discrimination of several emotions based on low-level physical characteristics. The emotion identification points, which reflect the amount of information required for stable recognition, were shortest for anger and happiness for both speech and music, but recognition took longer to stabilize for music vs speech. This, in turn, suggests that acoustic cues that develop over time also play a role for emotion inferences (especially for music). Finally, acoustic cue patterns were positively correlated between speech and music, suggesting a shared acoustic code for expressing emotions.
Skip Nav Destination
Article navigation
May 28 2019
The time course of emotion recognition in speech and music
Henrik Nordström;
Henrik Nordström
a)
Department of Psychology, Stockholm University
, 106 91 Stockholm, Sweden
Search for other works by this author on:
Petri Laukka
Petri Laukka
Department of Psychology, Stockholm University
, 106 91 Stockholm, Sweden
Search for other works by this author on:
a)
Electronic mail: henrik.nordstrom@psychology.su.se
J. Acoust. Soc. Am. 145, 3058–3074 (2019)
Article history
Received:
July 02 2018
Accepted:
April 25 2019
Citation
Henrik Nordström, Petri Laukka; The time course of emotion recognition in speech and music. J. Acoust. Soc. Am. 1 May 2019; 145 (5): 3058–3074. https://doi.org/10.1121/1.5108601
Download citation file:
Pay-Per-View Access
$40.00
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Citing articles via
Vowel signatures in emotional interjections and nonlinguistic vocalizations expressing pain, disgust, and joy across languages
Maïa Ponsonnet, Christophe Coupé, et al.
The alveolar trill is perceived as jagged/rough by speakers of different languages
Aleksandra Ćwiek, Rémi Anselme, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Related Content
Processing of emotional sounds conveyed by the human voice and musical instruments
J Acoust Soc Am (September 2018)
Oscillatory brain activity in response to emotional sounds in musicians and non-musicians
J Acoust Soc Am (May 2017)
Cell-phone vs microphone recordings: Judging emotion in the voice
J. Acoust. Soc. Am. (September 2017)
The contribution of phonation type to the perception of vocal emotions in German: An articulatory synthesis study
J. Acoust. Soc. Am. (March 2015)
Acoustic profiles of distinct emotional expressions in laughter
J. Acoust. Soc. Am. (July 2009)