We propose a method for co-registrating speech articulatory/acoustic data from two modalities that provide complementary advantages. Electromagnetic Articulography (EMA) provides high temporal resolution (100 samples/second in WAVE system) and flesh-point tracking, while real-time Magnetic Resonance Imaging, rtMRI, (23 frames/second) offers a complete midsagittal view of the vocal tract, including articulated structures and the articulatory environment. Co-registration was achieved through iterative alignment in the acoustic and articulatory domains. Acoustic signals were aligned temporally using Dynamic Time Warping, while articulatory signals were aligned variously by minimization of mean total error between articulatometry data and estimated corresponding flesh points and by using mutual information derived from articulatory parameters for each sentence. We demonstrate our method on a subset of the TIMIT corpus elicited from a male and a female speaker of American English, and illustrate the benefits of co-registered multi-modal data in the study of liquid and fricative consonant production in rapid speech. [Supported by NIH and NSF grants.]
Skip Nav Destination
,
,
,
Article navigation
September 2012
Meeting abstract. No PDF available.
September 01 2012
Co-registration of articulographic and real-time magnetic resonance imaging data for multimodal analysis of rapid speech
Jangwon Kim;
Jangwon Kim
Electrical Engineering, University of Southern California, 3740 McClintock Avenue, Los Angeles, CA [email protected]
Search for other works by this author on:
Adam Lammert;
Adam Lammert
Computer Science, University of Southern California, Los Angeles, CA
Search for other works by this author on:
Michael Proctor;
Michael Proctor
Electrical Engineering, University of Southern California, Los Angeles, CA
Search for other works by this author on:
Shrikanth Narayanan
Shrikanth Narayanan
Electrical Engineering, University of Southern California, Los Angeles, CA
Search for other works by this author on:
Jangwon Kim
Adam Lammert
Michael Proctor
Shrikanth Narayanan
Electrical Engineering, University of Southern California, 3740 McClintock Avenue, Los Angeles, CA [email protected]
J. Acoust. Soc. Am. 132, 2090 (2012)
Citation
Jangwon Kim, Adam Lammert, Michael Proctor, Shrikanth Narayanan; Co-registration of articulographic and real-time magnetic resonance imaging data for multimodal analysis of rapid speech. J. Acoust. Soc. Am. 1 September 2012; 132 (3_Supplement): 2090. https://doi.org/10.1121/1.4755722
Download citation file:
91
Views
Citing articles via
A survey of sound source localization with deep learning methods
Pierre-Amaury Grumiaux, Srđan Kitić, et al.
I can't hear you without my glasses
Tessa Bent
Related Content
Co-registration of speech production datasets from electromagnetic articulography and real-time magnetic resonance imaging
J. Acoust. Soc. Am. (January 2014)
Multimodal acquisition of articulatory data: Geometrical and temporal registration
J. Acoust. Soc. Am. (February 2016)
Estimating the control parameters of an articulatory model from electromagnetic articulograph data
J. Acoust. Soc. Am. (May 2011)
Real-time magnetic resonance imaging and electromagnetic articulography database for speech production research (TC)
J. Acoust. Soc. Am. (September 2014)
A study of data normalization measured by an electro-magnetic articulograph
J. Acoust. Soc. Am. (May 2013)