Much past effort, including our own, has gone into the design of high‐resolution time‐frequency representations. However, as found in most data‐trained automatic classification applications, high resolution can significantly degrade performance. This behavior can be understood from the standpoint of complexity: The higher the resolution and complexity of the time‐frequency representation, the larger the size of the training data needed for an accurate classifier. For most applications, where many regions in time and frequency are not salient to the classification task at hand, lowering resolution in these regions should actually improve classifier performance and generality. A new time‐frequency approach, called ‘‘class‐dependent kernels,’’ which selectively smoothes in time and frequency, has been developed. This approach can be summarized by considering the points of a discrete auto‐ambiguity function, which is the two‐dimensional Fourier transform of a time‐frequency representation, as a set of features, which can be ranked in terms of saliency to the particular classification task. Keeping and classifying based on only a small number of these ambiguity plane points corresponds to a flexible and data‐adaptive smoothing of the corresponding time‐frequency plane. This approach will be shown to be useful for a wide range of acoustic and vibration problems including monitoring acoustic emissions in titanium removal and as an adjunct to cepstra in speaker‐independent phoneme recognition.
Skip Nav Destination
Article navigation
October 1999
Meeting abstract. No PDF available.
October 01 1999
Adapting time‐frequency responses to speech and acoustic classification tasks Free
Les E. Atlas
Les E. Atlas
Dept. of Elec. Eng., Univ. of Washington, Box 352500, Seattle, WA 98195‐2500, [email protected]
Search for other works by this author on:
Les E. Atlas
Dept. of Elec. Eng., Univ. of Washington, Box 352500, Seattle, WA 98195‐2500, [email protected]
J. Acoust. Soc. Am. 106, 2156 (1999)
Citation
Les E. Atlas; Adapting time‐frequency responses to speech and acoustic classification tasks. J. Acoust. Soc. Am. 1 October 1999; 106 (4_Supplement): 2156. https://doi.org/10.1121/1.427386
Download citation file:
57
Views
Citing articles via
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Drawer-like tunable ventilated sound barrier
Yong Ge, Yi-jun Guan, et al.
Related Content
Ultrasonic analysis of breast tissue for pathology classification
J. Acoust. Soc. Am. (October 2011)
Back–front classification of English vowels using a cepstrum‐to‐formant model
J. Acoust. Soc. Am. (November 1995)
Timbre of Kiais: Analysis and synthesis
J. Acoust. Soc. Am. (August 2005)
Characterization of underwater explosions by cepstral analysis, modeling and inversion
J. Acoust. Soc. Am. (May 2001)
Cepstrogram and its application to speech analysis
J. Acoust. Soc. Am. (November 1994)