A computational model of auditory analysis is described that is inspired by psychoacoustical and neurophysiological findings in early and central stages of the auditory system. The model provides a unified multiresolution representation of the spectral and temporal features likely critical in the perception of sound. Simplified, more specifically tailored versions of this model have already been validated by successful application in the assessment of speech intelligibility [Elhilali et al., Speech Commun. 41(2-3), 331–348 (2003); Chi et al., J. Acoust. Soc. Am. 106, 2719–2732 (1999)] and in explaining the perception of monaural phase sensitivity [R. Carlyon and S. Shamma, J. Acoust. Soc. Am. 114, 333–348 (2003)]. Here we provide a more complete mathematical formulation of the model, illustrating how complex signals are transformed through various stages of the model, and relating it to comparable existing models of auditory processing. Furthermore, we outline several reconstruction algorithms to resynthesize the sound from the model output so as to evaluate the fidelity of the representation and contribution of different features and cues to the sound percept.
Skip Nav Destination
,
,
Article navigation
August 2005
August 01 2005
Multiresolution spectrotemporal analysis of complex sounds
Taishih Chi;
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742
Search for other works by this author on:
Powen Ru;
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742
Search for other works by this author on:
Shihab A. Shamma
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742
Search for other works by this author on:
Taishih Chi
Powen Ru
Shihab A. Shamma
Center for Auditory and Acoustics Research, Institute for Systems Research Electrical and Computer Engineering Department,
University of Maryland
, College Park, Maryland 20742a)
Present address: Department of Communication Engineering, National Chiao Tung University, Hsinchu, Taiwan, Republic of China
b)
Present address: Cybernetics InfoTech Inc.
c)
Electronic mail: [email protected]
J. Acoust. Soc. Am. 118, 887–906 (2005)
Article history
Received:
June 22 2004
Accepted:
May 12 2005
Citation
Taishih Chi, Powen Ru, Shihab A. Shamma; Multiresolution spectrotemporal analysis of complex sounds. J. Acoust. Soc. Am. 1 August 2005; 118 (2): 887–906. https://doi.org/10.1121/1.1945807
Download citation file:
Pay-Per-View Access
$40.00
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Citing articles via
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
Related Content
Auditory perception through the cortical analysis of spectrotemporal modulations.
J. Acoust. Soc. Am. (April 2009)
Auditory “bubbles”: Efficient classification of the spectrotemporal modulations essential for speech intelligibility
J. Acoust. Soc. Am. (August 2016)
Topographical distribution of spectrotemporal receptive field properties in the bat primary auditory cortex
J. Acoust. Soc. Am. (April 2022)
Forward masking of spectrotemporal modulation detection
J. Acoust. Soc. Am. (February 2022)
Learning spectro-temporal representations of complex sounds with parameterized neural networks
J. Acoust. Soc. Am. (July 2021)