Previous research suggests that many speech sounds are perceived in discrete categorical units. For example, perceivers typically identify acoustic stimuli from a continuum that ranges from one syllable to another (e.g., /va/-to-/ba/) exclusively as one syllable with a sharp change in which syllable the stimuli are identified as in the middle of the continuum. Further, pairs of continuum stimuli are typically easier to discriminate if they span this sharp change in stimulus identification. Walden et al. (1987) investigated whether visual (lipread) speech is perceived categorically using a continuum of digitally morphed mouths, but failed to find any strong evidence for categorical perception. In the current investigation, a technique developed by Baart and Vroomen (2010) was used to create a /va/-to-/ba/ continuum of visual test-stimuli. A video of a talker articulating /va/ was digitally superimposed over a video of the same talker articulating /ba/. Opacity of the /va/ video was then adjusted in 10% increments to create an 11-step continuum (from 0% to 100%). Identification and discrimination functions indicate that the visual continuum stimuli were categorically perceived as /va/ or /ba/. The results suggest that visual speech, like auditory speech, can be categorically perceived. Implications for gestural theories of speech will be discussed.
Skip Nav Destination
,
Article navigation
April 2016
Meeting abstract. No PDF available.
April 01 2016
Categorical perception of visual speech information Free
Chase Weinholtz;
Chase Weinholtz
Pitzer College, 1050 North Mills Ave., Pitzer College 835, Claremont, CA 91711, [email protected]
Search for other works by this author on:
James W. Dias
James W. Dias
Psych., Pomona College, Riverside, CA
Search for other works by this author on:
Chase Weinholtz
James W. Dias
Pitzer College, 1050 North Mills Ave., Pitzer College 835, Claremont, CA 91711, [email protected]
J. Acoust. Soc. Am. 139, 2018 (2016)
Citation
Chase Weinholtz, James W. Dias; Categorical perception of visual speech information. J. Acoust. Soc. Am. 1 April 2016; 139 (4_Supplement): 2018. https://doi.org/10.1121/1.4949950
Download citation file:
Citing articles via
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Variation in global and intonational pitch settings among black and white speakers of Southern American English
Aini Li, Ruaridh Purse, et al.
Related Content
Selective adaptation of crossmodal speech information is not the result of higher-level stimulus associations
J. Acoust. Soc. Am. (April 2016)
Phonetic recalibration of visual and auditory speech by visual sentences
J. Acoust. Soc. Am. (September 2018)
Acoustic noise and vision differentially warp the auditory categorization of speech
J. Acoust. Soc. Am. (July 2019)
Lexically guided retuning of visual phonetic categories
J. Acoust. Soc. Am. (July 2013)
Binding and unbinding the auditory and visual streams in the McGurk effect
J. Acoust. Soc. Am. (August 2012)