Code-switched Language Models Using Dual RNNs and Same-Source Pretraining

This work focuses on building language models (LMs) for code-switched text. We propose two techniques that significantly improve these LMs: 1) A novel recurrent neural network unit with dual components that focus on each language in the code-switched text separately 2) Pretraining the LM using synthetic text from a generative model estimated using the training data. We demonstrate the effectiveness of our proposed techniques by reporting perplexities on a Mandarin-English task and derive significant reductions in perplexity.

[1]  Preethi Jyothi,et al.  Dual Language Models for Code Mixed Speech Recognition , 2018, INTERSPEECH.

[2]  Monojit Choudhury,et al.  Curriculum Design for Code-switching: Experiments with Language Identification and Language Modeling with Deep Neural Networks , 2017, ICON.

[3]  Ying Li,et al.  Improved mixed language speech recognition using asymmetric acoustic model and language model with code-switch inversion constraints , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[4]  E. Chng,et al.  An Analysis of a Mandarin-English Code-switching Speech Corpus : SEAME , 2010 .

[5]  Ngoc Thang Vu,et al.  Combining recurrent neural networks and factored language models during decoding of code-Switching speech , 2014, INTERSPEECH.

[6]  Haizhou Li,et al.  Recurrent neural network language modeling for code switching conversational speech , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[7]  Ngoc Thang Vu,et al.  Challenges of Computational Processing of Code-Switching , 2016, CodeSwitch@EMNLP.

[8]  Tan Lee,et al.  Detection of language boundary in code-switching utterances by bi-phone probabilities , 2004, 2004 International Symposium on Chinese Spoken Language Processing.

[9]  Monojit Choudhury,et al.  Language Modeling for Code-Mixing: The Role of Linguistic Theory based Synthetic Data , 2018, ACL.

[10]  Ngoc Thang Vu,et al.  Syntactic and Semantic Features For Code-Switching Factored Language Models , 2015, IEEE/ACM Transactions on Audio, Speech, and Language Processing.

[11]  Pascale Fung,et al.  Language Modeling with Functional Head Constraint for Code Switching Speech Recognition , 2014, EMNLP.

[12]  Sumeet Singh,et al.  Language Informed Modeling of Code-Switched Text , 2018, CodeSwitch@ACL.

[13]  A. Backus Code-switching in conversation: Language, interaction and identity , 2000 .

[14]  Ying Li,et al.  Asymmetric acoustic modeling of mixed language speech , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[15]  Lantao Yu,et al.  SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.

[16]  Samy Bengio,et al.  Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.

[17]  Hervé Bourlard,et al.  Language dependent universal phoneme posterior estimation for mixed language speech recognition , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[18]  Robert L. Mercer,et al.  Class-Based n-gram Models of Natural Language , 1992, CL.

[19]  Lin-Shan Lee,et al.  An integrated framework for transcribing Mandarin-English code-mixed lectures with improved acoustic and language modeling , 2010, 2010 7th International Symposium on Chinese Spoken Language Processing.

[20]  Haizhou Li,et al.  A first speech recognition system for Mandarin-English code-switch conversational speech , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[21]  Lukás Burget,et al.  Recurrent neural network based language model , 2010, INTERSPEECH.