The efficacy of audio-visual interactions in speech perception comes from two kinds of factors. First, at the information level, there is some “complementarity” of audition and vision: It seems that some speech features, mainly concerned with manner of articulation, are best transmitted by the audio channel, while some other features, mostly describing place of articulation, are best transmitted by the video channel. Second, at the information processing level, there is some “synergy” between audition and vision: The audio-visual global identification scores in a number of different tasks involving acoustic noise are generally greater than both the auditory-alone and the visual-alone scores. However, these two properties have been generally demonstrated until now in rather global terms. In the present work, audio-visual interactions at the feature level are studied for French oral vowels which contrast three series, namely front unrounded, front rounded, and back rounded vowels. A set of experiments on the auditory, visual, and audio-visual identification of vowels embedded in various amounts of noise demonstrate that complementarity and synergy in bimodal speech appear to hold for a bundle of individual phonetic features describing place contrasts in oral vowels. At the information level (complementarity), in the audio channel the height feature is the most robust, backness the second most robust one, and rounding the least, while in the video channel rounding is better than height, and backness is almost invisible. At the information processing (synergy) level, transmitted information scores show that all individual features are better transmitted with the ear and the eye together than with each sensor individually.
Skip Nav Destination
Article navigation
June 1998
June 01 1998
Complementarity and synergy in bimodal speech: Auditory, visual, and audio-visual identification of French oral vowels in noise
Jordi Robert-Ribes;
Jordi Robert-Ribes
Institut de la Communication Parlée CNRS UPRESA 5009/INPG-U. Stendhal ICP, INPG, 46 Av. Félix-Viallet, 38031 Grenoble Cedex 1, France
Search for other works by this author on:
Jean-Luc Schwartz;
Jean-Luc Schwartz
Institut de la Communication Parlée CNRS UPRESA 5009/INPG-U. Stendhal ICP, INPG, 46 Av. Félix-Viallet, 38031 Grenoble Cedex 1, France
Search for other works by this author on:
Tahar Lallouache;
Tahar Lallouache
Institut de la Communication Parlée CNRS UPRESA 5009/INPG-U. Stendhal ICP, INPG, 46 Av. Félix-Viallet, 38031 Grenoble Cedex 1, France
Search for other works by this author on:
Pierre Escudier
Pierre Escudier
Institut de la Communication Parlée CNRS UPRESA 5009/INPG-U. Stendhal ICP, INPG, 46 Av. Félix-Viallet, 38031 Grenoble Cedex 1, France
Search for other works by this author on:
J. Acoust. Soc. Am. 103, 3677–3689 (1998)
Article history
Received:
July 31 1996
Accepted:
February 24 1998
Citation
Jordi Robert-Ribes, Jean-Luc Schwartz, Tahar Lallouache, Pierre Escudier; Complementarity and synergy in bimodal speech: Auditory, visual, and audio-visual identification of French oral vowels in noise. J. Acoust. Soc. Am. 1 June 1998; 103 (6): 3677–3689. https://doi.org/10.1121/1.423069
Download citation file:
Sign in
Don't already have an account? Register
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Sign in via your Institution
Sign in via your InstitutionPay-Per-View Access
$40.00