Models have long been used to understand the relation of anatomical structure and articulatory movement to the acoustics and perception of speech. Realized as speech synthesizers or artificial talkers, such models simplify and emulate the speech production system. One type of simplification is to view speech production as a set of simultaneously imposed modulations of the airway system. Specifically, the vibratory motion of the vocal folds modulates the glottal airspace, while slower movements of the tongue, jaw, lips, and velum modulate the shape of the pharyngeal and oral cavities, and coupling to the nasal system. The precise timing of these modulations produces an acoustic wave from which listeners extract phonetic and talker-specific information. The first aim of the presentation will be to review two historical models of speech production that exemplify a system in which structure is modulated with movement to produce intelligible speech. The second aim is to describe theoretical aspects of a computational model that allows for simulation of speech based on precise spatio-temporal modulations of an airway structure. The result is a type of artificial talker that can be used to study various aspects of how sound is generated by a speaker and perceived by a listener.
Skip Nav Destination
Article navigation
October 2016
Meeting abstract. No PDF available.
October 01 2016
The role of artificial speech in understanding the acoustic characteristics of spoken communication
Brad H. Story
Brad H. Story
Speech, Lang., and Hearing Sci., Univ. of Arizona, 1131 E. 2nd St., P.O. Box 210071, Tucson, AZ 85721, bstory@email.arizona.edu
Search for other works by this author on:
J. Acoust. Soc. Am. 140, 3316 (2016)
Citation
Brad H. Story; The role of artificial speech in understanding the acoustic characteristics of spoken communication. J. Acoust. Soc. Am. 1 October 2016; 140 (4_Supplement): 3316. https://doi.org/10.1121/1.4970564
Download citation file:
41
Views
Citing articles via
Vowel signatures in emotional interjections and nonlinguistic vocalizations expressing pain, disgust, and joy across languages
Maïa Ponsonnet, Christophe Coupé, et al.
The alveolar trill is perceived as jagged/rough by speakers of different languages
Aleksandra Ćwiek, Rémi Anselme, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Related Content
Articulatory comparison of spoken and sung vowels based on MRI
J Acoust Soc Am (May 2008)
A three-dimensional articulatory model of the velum and nasopharyngeal wall based on MRI and CT data
J. Acoust. Soc. Am. (April 2008)
High-dimensional and low-dimensional models of tissue and fluid movement in airways: How will they converge?
J Acoust Soc Am (May 2013)
A study on transvelar coupling for non-nasalized sounds
J. Acoust. Soc. Am. (January 2016)
Superior lateral pharyngeal wall movements in speech
J Acoust Soc Am (September 1986)