Real-time magnetic resonance imaging and electromagnetic articulography database for speech production research (TC).

USC-TIMIT is an extensive database of multimodal speech production data, developed to complement existing resources available to the speech research community and with the intention of being continuously refined and augmented. The database currently includes real-time magnetic resonance imaging data from five male and five female speakers of American English. Electromagnetic articulography data have also been presently collected from four of these speakers. The two modalities were recorded in two independent sessions while the subjects produced the same 460 sentence corpus used previously in the MOCHA-TIMIT database. In both cases the audio signal was recorded and synchronized with the articulatory data. The database and companion software are freely available to the research community.

[1]  M Stone,et al.  Comparison of speech production in upright and supine position. , 2007, The Journal of the Acoustical Society of America.

[2]  Shrikanth Narayanan,et al.  Synchronized and noise-robust audio recordings during realtime magnetic resonance imaging scans. , 2006, The Journal of the Acoustical Society of America.

[3]  Shrikanth S. Narayanan,et al.  Timing effects of syllable structure and stress on nasals: A real-time MRI examination , 2009, J. Phonetics.

[4]  Shrikanth Narayanan,et al.  An approach to real-time magnetic resonance imaging for speech production. , 2003, The Journal of the Acoustical Society of America.

[5]  Yoon-Chul Kim,et al.  Seeing speech: Capturing vocal tract shaping using real-time magnetic resonance imaging [Exploratory DSP] , 2008, IEEE Signal Processing Magazine.

[6]  Raymond D. Kent,et al.  X‐ray microbeam speech production database , 1990 .

[7]  Prasanta Kumar Ghosh,et al.  Co-registration of speech production datasets from electromagnetic articulography and real-time magnetic resonance imaging. , 2014, The Journal of the Acoustical Society of America.

[8]  J.M. Santos,et al.  Flexible real-time magnetic resonance imaging framework , 2004, The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[9]  Athanasios Katsamanis,et al.  Rapid semi-automatic segmentation of real-time magnetic resonance images for parametric vocal tract analysis , 2010, INTERSPEECH.

[10]  Jens Frahm,et al.  Real‐time MRI of speaking at a resolution of 33 ms: Undersampled radial FLASH with nonlinear inverse reconstruction , 2013, Magnetic resonance in medicine.

[11]  Shrikanth S. Narayanan,et al.  Flexible retrospective selection of temporal resolution in real‐time speech MRI using a golden‐ratio spiral view order , 2011, Magnetic resonance in medicine.

[12]  Zhi-Pei Liang,et al.  Dynamic imaging of speech and swallowing with MRI , 2009, 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[13]  Shrikanth S. Narayanan,et al.  Region Segmentation in the Frequency Domain Applied to Upper Airway Real-Time Magnetic Resonance Images , 2009, IEEE Transactions on Medical Imaging.

[14]  M H Cohen,et al.  Electromagnetic midsagittal articulometer systems for transducing speech articulatory movements. , 1992, The Journal of the Acoustical Society of America.

[15]  A. Macovski,et al.  Selection of a convolution function for Fourier inversion using gridding [computerised tomography application]. , 1991, IEEE transactions on medical imaging.

[16]  R. Boubertakh,et al.  Towards clinical assessment of velopharyngeal closure using MRI: evaluation of real-time MRI sequences at 1.5 and 3 T. , 2012, The British journal of radiology.

[17]  Shrikanth Narayanan,et al.  Paralinguistic mechanisms of production in human "beatboxing": a real-time magnetic resonance imaging study. , 2013, The Journal of the Acoustical Society of America.

[18]  Jens Frahm,et al.  Real‐time magnetic resonance imaging of normal swallowing , 2012, Journal of magnetic resonance imaging : JMRI.