Dividing the acoustic repertoires of animals into biologically relevant categories presents a widespread problem in the study of animal sound communication, essential to any comparison of repertoires between contexts, individuals, populations, or species. Automated procedures allow rapid, repeatable, and objective categorization, but often perform poorly at detecting biologically meaningful sound classes. Arguably this is because many automated methods fail to address the nonlinearities of animal sound perception. We present a new method of categorization that incorporates dynamic time-warping and an adaptive resonance theory (ART) neural network. This method was tested on 104 randomly chosen whistle contours from four captive bottlenose dolphins (Tursiops truncatus), as well as 50 frequency contours extracted from calls of transient killer whales (Orcinus orca). The dolphin data included known biologically meaningful categories in the form of 42 stereotyped whistles produced when each individual was isolated from its group. The automated procedure correctly grouped all but two stereotyped whistles into separate categories, thus performing as well as human observers. The categorization of killer whale calls largely corresponded to visual and aural categorizations by other researchers. These results suggest that this methodology provides a repeatable and objective means of dividing bioacoustic signals into biologically meaningful categories.
Skip Nav Destination
Article navigation
January 2006
January 01 2006
Automated categorization of bioacoustic signals: Avoiding perceptual pitfalls
Volker B. Deecke;
Volker B. Deecke
Marine Mammal Research Unit,
University of British Columbia
, 2202 Main Mall, Vancouver, BC V6T 1Z4 Canada and Cetacean Research Lab, Vancouver Aquarium Marine Science Centre
, P.O. Box 3232, Vancouver, BC V6B 3X8, Canada
Search for other works by this author on:
Vincent M. Janik
Vincent M. Janik
Sea Mammal Research Unit, Gatty Marine Laboratory,
University of St. Andrews
, Fife KY16 8LB, United Kingdom and Centre for Social Learning and Cognitive Evolution, School of Biology, University of St. Andrews
, Fife KY16 9TS, United Kingdom
Search for other works by this author on:
J. Acoust. Soc. Am. 119, 645–653 (2006)
Article history
Received:
June 06 2005
Accepted:
October 13 2005
Citation
Volker B. Deecke, Vincent M. Janik; Automated categorization of bioacoustic signals: Avoiding perceptual pitfalls. J. Acoust. Soc. Am. 1 January 2006; 119 (1): 645–653. https://doi.org/10.1121/1.2139067
Download citation file:
Pay-Per-View Access
$40.00
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Citing articles via
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Rapid detection of fish calls within diverse coral reef soundscapes using a convolutional neural network
Seth McCammon, Nathan Formel, et al.
Related Content
Automated categorization of bioacoustic signals: Avoiding perceptual pitfalls
J Acoust Soc Am (April 2005)
What’s in a voice? Cues used by dolphins in individual recognition of signature whistles
J Acoust Soc Am (September 2005)
Deep metric learning for bioacoustic classification: Overcoming training data scarcity using dynamic triplet loss
J. Acoust. Soc. Am. (July 2019)
Pitch- and spectral-based dynamic time warping methods for comparing field recordings of harmonic avian vocalizations
J. Acoust. Soc. Am. (August 2013)
Integration over song classification replicates: Song variant analysis in the hihi
J. Acoust. Soc. Am. (May 2015)