The aim of this research is to investigate machine learning techniques for emotion recognition for audio such as human speech and music. Emotion recognition technology has a history in commercial applications to enhance personalized music recommendations, mood-based ambience detection and interpret human interactions or human-machine interaction such as job interviews, caller-agent calls, streaming videos, and music platforms such as Spotify. Moreover, enhancing these algorithms can significantly benefit individuals on the autism spectrum by promoting accurate and practical support. In this study, we employed a combination of techniques to build a machine learning approach for emotion recognition. By adjusting the audio features and given test and train data, we aimed to identify and enhance the relationships between audio perception and its features. This approach seeks to eventually improve the accuracy and applicability of emotion recognition systems, contributing to the ongoing development and promotion of this technology in various domains.
Skip Nav Destination
Article navigation
October 2024
October 01 2024
Advancing emotion recognition of audio: Testing and verifying machine learning techniques for speech and music
Diana Gomez-Fonseca
Diana Gomez-Fonseca
None, 4539 Basswood Dr., Erie, PA 16506, [email protected]
Search for other works by this author on:
Diana Gomez-Fonseca
None, 4539 Basswood Dr., Erie, PA 16506, [email protected]
J. Acoust. Soc. Am. 156, A126 (2024)
Citation
Diana Gomez-Fonseca; Advancing emotion recognition of audio: Testing and verifying machine learning techniques for speech and music. J. Acoust. Soc. Am. 1 October 2024; 156 (4_Supplement): A126. https://doi.org/10.1121/10.0035340
Download citation file:
27
Views
Citing articles via
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
Related Content
Multimodal emotion perception: Influences of autism spectrum disorder and autism-like traits
J. Acoust. Soc. Am. (October 2020)
Emotion recognition in speech and music of Cantonese children with autism spectrum disorder
J. Acoust. Soc. Am. (April 2022)
Acoustics and emotion in tonal and non-tonal languages: Findings from individuals with typical hearing and with cochlear implants
J. Acoust. Soc. Am. (September 2018)
Emotion equalization app: The effectiveness of dynamic music therapy approaches
J. Acoust. Soc. Am. (March 2024)
Exploring the therapeutic effects of emotion equalization app during daily walking activities
J. Acoust. Soc. Am. (October 2024)