Recurrent Neural Network based language modeling with controllable external Memory
暂无分享,去创建一个
[1] Wei-Chen Cheng,et al. Language modeling with sum-product networks , 2014, INTERSPEECH.
[2] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[3] Christof Monz,et al. Recurrent Memory Networks for Language Modeling , 2016, NAACL.
[4] Tomas Mikolov,et al. Inferring Algorithmic Patterns with Stack-Augmented Recurrent Nets , 2015, NIPS.
[5] Tony Robinson,et al. Scaling recurrent neural network language models , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[6] Ahmad Emami,et al. Exact training of a neural syntactic language model , 2004, 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing.
[7] Mark J. F. Gales,et al. Investigation of back-off based interpolation between recurrent neural network and n-gram language models , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).
[8] Steve Renals,et al. Feed forward pre-training for recurrent neural network language models , 2014, INTERSPEECH.
[9] Jason Weston,et al. End-To-End Memory Networks , 2015, NIPS.
[10] Hermann Ney,et al. On efficient training of word classes and their application to recurrent neural network language models , 2015, INTERSPEECH.
[11] Alex Graves,et al. Neural Turing Machines , 2014, ArXiv.
[12] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[13] Lukás Burget,et al. Extensions of recurrent neural network language model , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[14] Hermann Ney,et al. LSTM Neural Networks for Language Modeling , 2012, INTERSPEECH.
[15] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[16] Inchul Song,et al. RNNDROP: A novel dropout for RNNS in ASR , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).
[17] Michael I. Jordan,et al. Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..
[18] Peter Kulchyski. and , 2015 .
[19] Hermann Ney,et al. rwthlm - the RWTH aachen university neural network language modeling toolkit , 2014, INTERSPEECH.
[20] Geoffrey E. Hinton,et al. Three new graphical models for statistical language modelling , 2007, ICML '07.
[21] Marc'Aurelio Ranzato,et al. Learning Longer Memory in Recurrent Neural Networks , 2014, ICLR.
[22] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[23] Mark J. F. Gales,et al. Recurrent neural network language model training with noise contrastive estimation for speech recognition , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[24] Hermann Ney,et al. Improved backing-off for M-gram language modeling , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.
[25] Geoffrey Zweig,et al. Context dependent recurrent neural network language model , 2012, 2012 IEEE Spoken Language Technology Workshop (SLT).
[26] Mark J. F. Gales,et al. Improving the training and evaluation efficiency of recurrent neural network language models , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[27] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[28] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[29] Tomas Mikolov,et al. RNNLM - Recurrent Neural Network Language Modeling Toolkit , 2011 .
[30] Christof Monz,et al. Recurrent Memory Network for Language Modeling , 2016, ArXiv.
[31] Kam-Fai Wong,et al. Recurrent Neural Networks with External Memory for Spoken Language Understanding , 2015, NLPCC.
[32] Mirella Lapata,et al. Long Short-Term Memory-Networks for Machine Reading , 2016, EMNLP.