Automatic speech recognition (ASR) systems are greatly demanded for customer service systems. With advanced interactive voice response systems, humans have more opportunities to have dialogues with computers. Existing dialogue systems process linguistic information, but do not process paralinguistic information. Therefore, computers are able to obtain less information during a human‐computer dialogue than a human can during a human‐human dialogue. This report describes a study of the estimation method of degree of speakers’ anger emotion using acoustic features and linguistic representation expressed in utterances during a natural dialogue. To record utterances expressing the users’ internal anger emotion, we set pseudo‐dialogues to induce irritation arising from discontentment with the ASR system performance and to induce exasperation against the operator while the user makes a complaint. A five‐scale subjective evaluation was conducted to mark each utterance with a score as the actual measurement of anger emotion. As a result of this, an emotional speech corpus was produced. We examine the acoustic features and features of linguistic representation of each utterance with reference to these anger score. Then we conduct experiments to estimate automatically the degree of anger emotion using parameters selected from those features.
Skip Nav Destination
,
,
Article navigation
November 2006
Meeting abstract. No PDF available.
November 01 2006
A method for estimating the degree of a speaker’s anger using acoustic features and linguistic representation Free
Yoshiko Arimoto;
Yoshiko Arimoto
Grad. School of Bionics, Comput. & Media Sci., Tokyo Univ. of Technol., 1401‐1 Katakura, Hachioji, Tokyo, 192‐0982, Japan
Search for other works by this author on:
Sumio Ohno;
Sumio Ohno
Tokyo Univ. of Tech., Hachioji, Tokyo, 192‐0982, Japan
Search for other works by this author on:
Hitoshi Iida
Hitoshi Iida
Tokyo Univ. of Tech., Hachioji, Tokyo, 192‐0982, Japan
Search for other works by this author on:
Yoshiko Arimoto
Sumio Ohno
Hitoshi Iida
Grad. School of Bionics, Comput. & Media Sci., Tokyo Univ. of Technol., 1401‐1 Katakura, Hachioji, Tokyo, 192‐0982, Japan
J. Acoust. Soc. Am. 120, 3044 (2006)
Citation
Yoshiko Arimoto, Sumio Ohno, Hitoshi Iida; A method for estimating the degree of a speaker’s anger using acoustic features and linguistic representation. J. Acoust. Soc. Am. 1 November 2006; 120 (5_Supplement): 3044. https://doi.org/10.1121/1.4787232
Download citation file:
69
Views
Citing articles via
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
Drawer-like tunable ventilated sound barrier
Yong Ge, Yi-jun Guan, et al.
Related Content
Study on voice quality parameters for anger degree estimation
J. Acoust. Soc. Am. (May 2008)
Automatic recognition of negative emotion in speech using support vector machine
J. Acoust. Soc. Am. (October 2016)
Realization of rhythmic dialogue on spoken dialogue system using paralinguistic information
J. Acoust. Soc. Am. (November 2006)
Adaptive interface for spoken dialog
J. Acoust. Soc. Am. (May 2002)
Concept‐to‐speech conversion for reply speech generation in a spoken dialogue system for road guidance and its prosodic control
J. Acoust. Soc. Am. (November 2006)