Classic accounts of the benefits of speechreading to speech recognition treat auditory and visual channels as independent sources of information that are integrated fairly early in the speech perception process. The primary question addressed in this study was whether visible movements of the speech articulators could be used to improve the detection of speech in noise, thus demonstrating an influence of speechreading on the ability to detect, rather than recognize, speech. In the first experiment, ten normal-hearing subjects detected the presence of three known spoken sentences in noise under three conditions: auditory-only auditory plus speechreading with a visually matched sentence and auditory plus speechreading with a visually unmatched sentence When the speechread sentence matched the target sentence, average detection thresholds improved by about 1.6 dB relative to the auditory condition. However, the amount of threshold reduction varied significantly for the three target sentences (from 0.8 to 2.2 dB). There was no difference in detection thresholds between the condition and the condition. In a second experiment, the effects of visually matched orthographic stimuli on detection thresholds was examined for the same three target sentences in six subjects who participated in the earlier experiment. When the orthographic stimuli were presented just prior to each trial, average detection thresholds improved by about 0.5 dB relative to the condition. However, unlike the condition, the detection improvement due to orthography was not dependent on the target sentence. Analyses of correlations between area of mouth opening and acoustic envelopes derived from selected spectral regions of each sentence (corresponding to the wide-band speech, and first, second, and third formant regions) suggested that threshold reduction may be determined by the degree of auditory-visual temporal coherence, especially between the area of lip opening and the envelope derived from mid- to high-frequency acoustic energy. Taken together, the data (for these sentences at least) suggest that visual cues derived from the dynamic movements of the fact during speech production interact with time-aligned auditory cues to enhance sensitivity in auditory detection. The amount of visual influence depends in part on the degree of correlation between acoustic envelopes and visible movement of the articulators.
Skip Nav Destination
Article navigation
September 2000
September 01 2000
The use of visible speech cues for improving auditory detection of spoken sentences
Ken W. Grant;
Ken W. Grant
Walter Reed Army Medical Center, Army Audiology and Speech Center, Washington, D.C. 20307-5001
Search for other works by this author on:
Philip-Franz Seitz
Philip-Franz Seitz
Walter Reed Army Medical Center, Army Audiology and Speech Center, Washington, D.C. 20307-5001
Search for other works by this author on:
J. Acoust. Soc. Am. 108, 1197–1208 (2000)
Article history
Received:
June 29 1999
Accepted:
June 12 2000
Citation
Ken W. Grant, Philip-Franz Seitz; The use of visible speech cues for improving auditory detection of spoken sentences. J. Acoust. Soc. Am. 1 September 2000; 108 (3): 1197–1208. https://doi.org/10.1121/1.1288668
Download citation file:
Sign in
Don't already have an account? Register
Sign In
You could not be signed in. Please check your credentials and make sure you have an active account and try again.
Pay-Per-View Access
$40.00
Citing articles via
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Short-time coherence between repeated room impulse response measurements
Karolina Prawda, Sebastian J. Schlecht, et al.
Efficient design of complex-valued neural networks with application to the classification of transient acoustic signals
Vlad S. Paul, Philip A. Nelson