Smith and Lewicki, Neural Comp. 17, 19–45, 2005a; Adv. Neural. Inf. Process. Syst. 17, 1289–1296, 2005b; Nature, 439, 7079, 2006, demonstrated that mammalian hearing follows an efficient coding principle (Barlow, Sensory Communications, 217–234, 1961; Atick, Network, 3(2), 213–251, 1992; Simoncelli and Olshausen, Ann. Rev. Neurosci., 24, 1193–1216, 2001; Laughlin and Sejnowski, ‘‘Communications in Neuronal Networks,’’ Science, 301, 1870–1874 2003). Auditory neurons efficiently code for natural sounds in the environment, maximizing information rate while minimizing coding cost (Shannon, Science, 270, 303–304, 1948). Applying the same analysis to speech coding suggests that speech acoustics are optimally adapted to the mammalian auditory code (Smith and Lewicki, Neural Comp. 17, 19–45, 2005a; Adv. Neural. Inf. Process. Syst. 17, 1289–1296, 2005b; Nature, 439, 7079, 2006). The present work applies this efficient coding theory to the problem of speech perception in individuals using cochlear implants (CI), for which there exist vast individual differences in speech perception and spectral resolution (Zeng et al., Auditory Prostheses and Electric Hearing, 20, 1–14, 2004). A machine‐learning method for CI filterbank design based on the efficient‐coding hypothesis is presented. Further, a pair of experiments to evaluate this approach using noise‐excited vocoder speech (Shannon et al., Bell Systems Technical Journal 27, 379–423, 623–656, 1995) is described. Participants’ recognition of continuous speech and isolated syllables is significantly more accurate for speech filtered through the theoretically‐motivated efficient‐coding filterbank relative to the standard cochleotopic filterbank, particularly for speech transients. These findings offer insight in CI design and provide behavioral evidence for efficient coding in human perception.
Skip Nav Destination
Article navigation
May 2006
Meeting abstract. No PDF available.
May 04 2006
A theoretical model of cochlear processing improves spectrally degraded speech perception
Evan C. Smith;
Evan C. Smith
Dept. Psych. & The Ctr. for the Neural Basis of Cognition, Carnegie Mellon Univ., Pittsburgh, PA 15213
Search for other works by this author on:
Lori L. Holt
Lori L. Holt
Dept. Psych. & The Ctr. for the Neural Basis of Cognition, Carnegie Mellon Univ., Pittsburgh, PA 15213
Search for other works by this author on:
J. Acoust. Soc. Am. 119, 3238 (2006)
Citation
Evan C. Smith, Lori L. Holt; A theoretical model of cochlear processing improves spectrally degraded speech perception. J. Acoust. Soc. Am. 1 May 2006; 119 (5_Supplement): 3238. https://doi.org/10.1121/1.4786001
Download citation file:
38
Views
Citing articles via
Vowel signatures in emotional interjections and nonlinguistic vocalizations expressing pain, disgust, and joy across languages
Maïa Ponsonnet, Christophe Coupé, et al.
The alveolar trill is perceived as jagged/rough by speakers of different languages
Aleksandra Ćwiek, Rémi Anselme, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Related Content
Efficient coding in human auditory perception
J. Acoust. Soc. Am. (September 2009)
Languages across the world are efficiently coded by the auditory system
J Acoust Soc Am (April 2015)
Analysis and design of gammatone signal models
J. Acoust. Soc. Am. (November 2009)
Statistical structure of speech sound classes is congruent with cochlear nucleus response properties
J Acoust Soc Am (November 2013)
Efficient coding of a xylophone sound using spikogram nonredundent coding
J Acoust Soc Am (May 2008)