A novel method of feature extraction for phoneme recognition in continuous speech is proposed that employs mutual information between acoustic features and phonemes. Various acoustic features are coded by the vector quantization (VQ) method and a method to discriminate phonemes by the effective combination of these VQ codes is developed. To construct an optimal algorithm for phoneme discrimination, entropy and mutual information, in addition to conditional probability, between phoneme labels and features are also taken into consideration. The effectiveness of each acoustic feature for describing the characteristics of the phoneme in a given environment is evaluated based on the mutual information. The LPC mel‐cepstrum. its pattern of temporal changes over frames, and power are used as acoustic features. Three experiments were conducted. The first was on the optimization of the frame labeling. The second was on the detection of the vowel using a sequence of frame labels. The third was on word discrimination. The effectiveness of the proposed method was verified by these experiments.
Skip Nav Destination
Article navigation
November 1988
August 13 2005
Phonetic feature extraction based on mutual information Free
Noriyuki Aoki;
Noriyuki Aoki
Department of Electrical Engineering, Waseda University, 3‐4‐1 Ohkubo, Shinjuku‐ku, Tokyo, 160 Japan
Search for other works by this author on:
Naoki Hosaka;
Naoki Hosaka
Department of Electrical Engineering, Waseda University, 3‐4‐1 Ohkubo, Shinjuku‐ku, Tokyo, 160 Japan
Search for other works by this author on:
Katsuhiko Shirai
Katsuhiko Shirai
Department of Electrical Engineering, Waseda University, 3‐4‐1 Ohkubo, Shinjuku‐ku, Tokyo, 160 Japan
Search for other works by this author on:
Noriyuki Aoki
Department of Electrical Engineering, Waseda University, 3‐4‐1 Ohkubo, Shinjuku‐ku, Tokyo, 160 Japan
Naoki Hosaka
Department of Electrical Engineering, Waseda University, 3‐4‐1 Ohkubo, Shinjuku‐ku, Tokyo, 160 Japan
Katsuhiko Shirai
Department of Electrical Engineering, Waseda University, 3‐4‐1 Ohkubo, Shinjuku‐ku, Tokyo, 160 Japan
J. Acoust. Soc. Am. 84, S211–S212 (1988)
Citation
Noriyuki Aoki, Naoki Hosaka, Katsuhiko Shirai; Phonetic feature extraction based on mutual information. J. Acoust. Soc. Am. 1 November 1988; 84 (S1): S211–S212. https://doi.org/10.1121/1.2026155
Download citation file:
80
Views
Citing articles via
Climatic and economic fluctuations revealed by decadal ocean soundscapes
Vanessa M. ZoBell, Natalie Posdaljian, et al.
Variation in global and intonational pitch settings among black and white speakers of Southern American English
Aini Li, Ruaridh Purse, et al.
Bioinspired flow-sensing capacitive microphone
Johar Pourghader, Weili Cui, et al.
Related Content
Speaker‐independent recognition of unvoiced plosives using a convex time pattern of a posteriori probability
J. Acoust. Soc. Am. (August 2005)
Across‐talker acoustic properties for place of articulation in nasal consonants
J. Acoust. Soc. Am. (August 2005)
A categorical factor analysis of vowel distribution based on the modified qualification theory
J. Acoust. Soc. Am. (August 2005)
Acoustic segmentation and phonetic classification in the SUMMIT system
J. Acoust. Soc. Am. (August 2005)
Word recognition using a two‐dimensional mel‐cepstrum in noisy environments
J. Acoust. Soc. Am. (August 2005)