While audiovisual interactions in speech perception have long been considered as automatic, recent data suggest that this is not the case. In a previous study, Nahorna et al. [(2012). J. Acoust. Soc. Am. 132, 1061–1077] showed that the McGurk effect is reduced by a previous incoherent audiovisual context. This was interpreted as showing the existence of an audiovisual binding stage controlling the fusion process. Incoherence would produce unbinding and decrease the weight of the visual input in fusion. The present paper explores the audiovisual binding system to characterize its dynamics. A first experiment assesses the dynamics of unbinding, and shows that it is rapid: An incoherent context less than 0.5 s long (typically one syllable) suffices to produce a maximal reduction in the McGurk effect. A second experiment tests the rebinding process, by presenting a short period of either coherent material or silence after the incoherent unbinding context. Coherence provides rebinding, with a recovery of the McGurk effect, while silence provides no rebinding and hence freezes the unbinding process. These experiments are interpreted in the framework of an audiovisual speech scene analysis process assessing the perceptual organization of an audiovisual speech input before decision takes place at a higher processing stage.
Skip Nav Destination
Article navigation
January 2015
January 01 2015
Audio-visual speech scene analysis: Characterization of the dynamics of unbinding and rebinding the McGurk effect Available to Purchase
Olha Nahorna;
Olha Nahorna
GIPSA-Lab, Speech and Cognition Department, UMR 5216, CNRS,
Grenoble University
, Grenoble, France
Search for other works by this author on:
Frédéric Berthommier;
Frédéric Berthommier
GIPSA-Lab, Speech and Cognition Department, UMR 5216, CNRS,
Grenoble University
, Grenoble, France
Search for other works by this author on:
Jean-Luc Schwartz
Jean-Luc Schwartz
a)
GIPSA-Lab, Speech and Cognition Department, UMR 5216, CNRS,
Grenoble University
, Grenoble, France
Search for other works by this author on:
Olha Nahorna
GIPSA-Lab, Speech and Cognition Department, UMR 5216, CNRS,
Grenoble University
, Grenoble, France
Frédéric Berthommier
GIPSA-Lab, Speech and Cognition Department, UMR 5216, CNRS,
Grenoble University
, Grenoble, France
Jean-Luc Schwartz
a)
GIPSA-Lab, Speech and Cognition Department, UMR 5216, CNRS,
Grenoble University
, Grenoble, France
a)
Author to whom correspondence should be addressed. Electronic mail: [email protected]
J. Acoust. Soc. Am. 137, 362–377 (2015)
Article history
Received:
November 19 2013
Accepted:
November 24 2014
Citation
Olha Nahorna, Frédéric Berthommier, Jean-Luc Schwartz; Audio-visual speech scene analysis: Characterization of the dynamics of unbinding and rebinding the McGurk effect. J. Acoust. Soc. Am. 1 January 2015; 137 (1): 362–377. https://doi.org/10.1121/1.4904536
Download citation file:
Pay-Per-View Access
$40.00
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Citing articles via
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Variation in global and intonational pitch settings among black and white speakers of Southern American English
Aini Li, Ruaridh Purse, et al.
Related Content
Binding and unbinding the auditory and visual streams in the McGurk effect
J. Acoust. Soc. Am. (August 2012)
Audiovisual speech perception: Moving beyond McGurk
J. Acoust. Soc. Am. (December 2022)
The early maximum likelihood estimation model of audiovisual integration in speech perception
J. Acoust. Soc. Am. (May 2015)
A reanalysis of McGurk data suggests that audiovisual fusion in speech perception is subject-dependent
J. Acoust. Soc. Am. (March 2010)
The influence of an inverted face on the McGurk effect
J. Acoust. Soc. Am. (May 1994)