Dolphins and whales use tonal whistles for communication, and it is known that frequency modulation encodes contextual information. An automated mathematical algorithm could characterize the frequency modulation of tonal calls for use with clustering and classification. Most automatic cetacean whistle processing techniques are based on peak or edge detection or require analyst assistance in verifying detections. An alternative paradigm is introduced using techniques of image processing. Frequency information is extracted as ridges in whistle spectrograms. Spectral ridges are the fundamental structure of tonal vocalizations, and ridge detection is a well-established image processing technique, easily applied to vocalization spectrograms. This paradigm is implemented as freely available matlab scripts, coined IPRiT (image processing ridge tracker). Its fidelity in the reconstruction of synthesized whistles is compared to another published whistle detection software package, silbido. Both algorithms are also applied to real-world recordings of bottlenose dolphin (Tursiops trunactus) signature whistles and tested for the ability to identify whistles belonging to different individuals. IPRiT gave higher fidelity and lower false detection than silbido with synthesized whistles, and reconstructed dolphin identity groups from signature whistles, whereas silbido could not. IPRiT appears to be superior to silbido for the extraction of the precise frequency variation of the whistle.
Skip Nav Destination
Article navigation
December 2013
December 01 2013
An image processing based paradigm for the extraction of tonal sounds in cetacean communications
Arik Kershenbaum;
Arik Kershenbaum
a)
National Institute for Mathematical and Biological Synthesis
, Knoxville, Tennessee 37996
Search for other works by this author on:
Marie A. Roch
Marie A. Roch
Department of Computer Science, San Diego State University
, San Diego, California 92182
Search for other works by this author on:
a)
Author to whom correspondence should be addressed. Electronic mail: arik@nimbios.org
J. Acoust. Soc. Am. 134, 4435–4445 (2013)
Article history
Received:
January 24 2013
Accepted:
October 14 2013
Citation
Arik Kershenbaum, Marie A. Roch; An image processing based paradigm for the extraction of tonal sounds in cetacean communications. J. Acoust. Soc. Am. 1 December 2013; 134 (6): 4435–4445. https://doi.org/10.1121/1.4828821
Download citation file:
Sign in
Don't already have an account? Register
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Could not validate captcha. Please try again.
Sign in via your Institution
Sign in via your InstitutionPay-Per-View Access
$40.00
Citing articles via
Related Content
Silbido profundo: An open source package for the use of deep learning to detect odontocete whistles
J. Acoust. Soc. Am. (December 2022)
A performance comparison of tonal detectors for low-frequency vocalizations of Antarctic blue whales
J. Acoust. Soc. Am. (January 2020)
Joint classification of whistles and echolocation clicks from odontocetes
J Acoust Soc Am (May 2013)
Joint classification of whistles and echolocation clicks from odontocetes
Proc. Mtgs. Acoust. (June 2013)
Auditory evoked potential measurements with cetaceans
J Acoust Soc Am (April 2005)