In this article, we present an approach for non native automatic speech recognition (ASR). We propose two methods to adapt existing ASR systems to the non-native accents. The first method is based on the modification of acoustic models through integration of acoustic models from the mother tong. The phonemes of the target language are pronounced in a similar manner to the native language of speakers. We propose to combine the models of confused phonemes so that the ASR system could recognize both concurrent pronounciations. The second method we propose is a refinment of the pronounciation error detection through the introduction of graphemic constraints. Indeed, non native speakers may rely on the writing of words in their uttering. Thus, the pronounctiation errors might depend on the characters composing the words. The average error rate reduction that we observed is (22.5%) relative for the sentence error rate, and 34.5% (relative) in word error rate.
[1]
P. Ladefoged,et al.
The sounds of the world's languages
,
1996
.
[2]
Jean Paul Haton,et al.
Fully Automated Non-Native Speech Recognition Using Confusion-Based Acoustic Model Integration and Graphemic Constraints
,
2006,
2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.
[3]
John J. Morgan,et al.
Making a Speech Recognizer Tolerate Non-native Speech through Gaussian Mixture Merging
,
2004
.
[4]
I. Lehiste,et al.
Principles and Methods for Historical Linguistics
,
1979
.
[5]
James R. Glass,et al.
Lexical modeling of non-native speech for automatic speech recognition
,
2000,
2000 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.00CH37100).
[6]
Stefan Schaden.
Generating Non - Native Pronuncia - tion Lexicons by Phonological Rule
,
2003
.