The goal of this research is to develop voice‐controlled wheelchairs that can be operated by inarticulate speech affected by severe cerebral palsy or quadriplegia, for instance. In this case, principal factors obstructing recognition performance are significant pronunciation variation caused by difficulty in stable articulation and bad influences of variety of noise in the real environment. To cope with the pronunciation variations, pronunciation lexicons that consist of multitemplates of reference patterns are utilized. The reference patterns are represented with subphonetic codes, which can describe variations of inarticulate speech more precisely than ordinary phonetic transcriptions. Pronunciation lexicons are generated by generalizing coded samples into compact representation of templates based on DP and data mining. For noise robustness, a voice activity detection method is investigated in order to circumvent friction of microphone, cough, etc. A sound source localization using a microphone array is also integrated in order to reject sounds from outside of the wheelchair. These methods are integrated into a system that can be mounted on a wheelchair. A usability test operated by a severe cerebral palsy patient in the real environment results in 95.8% accuracy within 1404 samples for the recognition of the five‐command set. [Work supported by MEXT.]
Skip Nav Destination
,
,
,
,
Article navigation
November 2006
Meeting abstract. No PDF available.
November 01 2006
Robust speech recognition for the control of wheelchairs by inarticulate speech of the severely disabled Free
Hiroaki Kojima;
Hiroaki Kojima
Natl. Inst. of Adv. Industrial Sci. and Technol. (AIST), Central 2, Umezono, Tsukuba, Ibaraki 305‐8568, Japan
Search for other works by this author on:
Akira Sasou;
Akira Sasou
Natl. Inst. of Adv. Industrial Sci. and Technol. (AIST), Central 2, Umezono, Tsukuba, Ibaraki 305‐8568, Japan
Search for other works by this author on:
Soo‐Young Suk;
Soo‐Young Suk
Natl. Inst. of Adv. Industrial Sci. and Technol. (AIST), Central 2, Umezono, Tsukuba, Ibaraki 305‐8568, Japan
Search for other works by this author on:
Shi‐Wook Lee;
Shi‐Wook Lee
Natl. Inst. of Adv. Industrial Sci. and Technol. (AIST), Central 2, Umezono, Tsukuba, Ibaraki 305‐8568, Japan
Search for other works by this author on:
Ken Sadohara
Ken Sadohara
Natl. Inst. of Adv. Industrial Sci. and Technol. (AIST), Central 2, Umezono, Tsukuba, Ibaraki 305‐8568, Japan
Search for other works by this author on:
Hiroaki Kojima
Akira Sasou
Soo‐Young Suk
Shi‐Wook Lee
Ken Sadohara
Natl. Inst. of Adv. Industrial Sci. and Technol. (AIST), Central 2, Umezono, Tsukuba, Ibaraki 305‐8568, Japan
J. Acoust. Soc. Am. 120, 3043 (2006)
Citation
Hiroaki Kojima, Akira Sasou, Soo‐Young Suk, Shi‐Wook Lee, Ken Sadohara; Robust speech recognition for the control of wheelchairs by inarticulate speech of the severely disabled. J. Acoust. Soc. Am. 1 November 2006; 120 (5_Supplement): 3043. https://doi.org/10.1121/1.4787223
Download citation file:
Citing articles via
Focality of sound source placement by higher (ninth) order ambisonics and perceptual effects of spectral reproduction errors
Nima Zargarnezhad, Bruno Mesquita, et al.
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
Drawer-like tunable ventilated sound barrier
Yong Ge, Yi-jun Guan, et al.
Related Content
Novel interface using a microphone array for wheelchair control
J. Acoust. Soc. Am. (November 2006)
Design of electrical wheelchair navigation for disabled patient using convolutional neural networks on Raspberry Pi 3
AIP Conf. Proc. (October 2020)
Mechanics in a wheelchair
Phys. Teach. (December 1999)
A Verbal Guidance System for Severe Disabled People
AIP Conf. Proc. (June 2008)
Voice activity detection with voice/nonvoice classification using reliable fundamental frequency
J. Acoust. Soc. Am. (November 2006)