Speech processing technologies, including speech recognition, synthesis, and coding are expected to play important roles in an advanced multimedia society with user‐friendly human–machine interfaces. Speech recognition systems include not only those that recognize messages but also those that recognize the identity of the speaker. This paper predicts future directions in speech processing. It describes the most important research problems and tries to forecast where progress will be made in the near future and what applications will become commonplace as a result of the increased capabilities. The most promising application area is telecommunications. To solve various fundamental problems, a unified approach across speech recognition, synthesis, and coding is indispensable. Handling the common phenomenon of voice individuality from different aspects are: research on speaker adaptation in speech recognition, automatic speaker verification, voice conversion in speech synthesis, and the problems of very low‐bit‐rate speech coding quality variation from speaker to speaker. The methods that essentially solve such problems should be based on a common mathematical model. Research on the mechanism of speech information processing in our human brains, that is, how meanings of speech are understood and how speech is produced, is also crucial for epoch‐making technological development in conversational speech understanding and natural speech synthesis.
Skip Nav Destination
Article navigation
May 1998
Meeting abstract. No PDF available.
May 01 1998
Future directions in speech information processing
Sadaoki Furui
Sadaoki Furui
Dept. of Comput. Sci., Tokyo Inst. of Technol., 2‐12‐1, Ookayama, Meguro‐ku, Tokyo, 152 Japan, [email protected]
Search for other works by this author on:
J. Acoust. Soc. Am. 103, 2747 (1998)
Citation
Sadaoki Furui; Future directions in speech information processing. J. Acoust. Soc. Am. 1 May 1998; 103 (5_Supplement): 2747. https://doi.org/10.1121/1.422797
Download citation file:
Citing articles via
All we know about anechoic chambers
Michael Vorländer
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Does sound symbolism need sound?: The role of articulatory movement in detecting iconicity between sound and meaning
Mutsumi Imai, Sotaro Kita, et al.
Related Content
Audio feature extraction using probability distribution function
AIP Conference Proceedings (May 2015)
Toward hidden Markov model‐based spontaneous speech synthesis
J Acoust Soc Am (November 2006)
On the role of spectral transition for speech perception
J Acoust Soc Am (October 1986)
Fifty years of progress in speech and speaker recognition
J Acoust Soc Am (October 2004)
Robust speech recognition for home environments using factorial hidden Markov models
J Acoust Soc Am (November 2006)