Larger-Context Language Modelling with Recurrent Neural Network
暂无分享,去创建一个
[1] Yoshua Bengio,et al. Neural Probabilistic Language Models , 2006 .
[2] Michael I. Jordan,et al. Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..
[3] Stanley F. Chen,et al. An empirical study of smoothing techniques for language modeling , 1999 .
[4] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[5] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[6] Geoffrey Zweig,et al. Context dependent recurrent neural network language model , 2012, 2012 IEEE Spoken Language Technology Workshop (SLT).
[7] Hermann Ney,et al. Improved backing-off for M-gram language modeling , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.
[8] Peter Wiemer-Hastings,et al. Latent semantic analysis , 2004, Annu. Rev. Inf. Sci. Technol..
[9] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[10] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[11] Joelle Pineau,et al. Hierarchical Neural Network Generative Models for Movie Dialogues , 2015, ArXiv.
[12] Lukás Burget,et al. Extensions of recurrent neural network language model , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[13] Derek Greene,et al. Practical solutions to the problem of diagonal dominance in kernel document clustering , 2006, ICML.
[14] Jason Weston,et al. End-To-End Memory Networks , 2015, NIPS.
[15] Chris Dyer,et al. Document Context Language Models , 2015, ICLR 2015.
[16] Jason Weston,et al. The Goldilocks Principle: Reading Children's Books with Explicit Memory Representations , 2015, ICLR.
[17] Cindy K. Chung,et al. The Psychological Functions of Function Words , 2007 .
[18] G. Miller. On knowing a word. , 1999, Annual review of psychology.
[19] Yoshua Bengio,et al. Gated Feedback Recurrent Neural Networks , 2015, ICML.
[20] Yoshua Bengio,et al. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches , 2014, SSST@EMNLP.
[21] R. Rosenfeld,et al. Two decades of statistical language modeling: where do we go from here? , 2000, Proceedings of the IEEE.
[22] Hermann Ney,et al. From Feedforward to Recurrent LSTM Neural Networks for Language Modeling , 2015, IEEE/ACM Transactions on Audio, Speech, and Language Processing.
[23] Dan Klein,et al. Feature-Rich Part-of-Speech Tagging with a Cyclic Dependency Network , 2003, NAACL.
[24] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[25] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[26] Philipp Koehn,et al. Scalable Modified Kneser-Ney Language Model Estimation , 2013, ACL.
[27] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[28] Jürgen Schmidhuber,et al. LSTM: A Search Space Odyssey , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[29] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[30] Phil Blunsom,et al. Pragmatic Neural Language Modelling in Machine Translation , 2014, NAACL.
[31] Mikel L. Forcada,et al. Recursive Hetero-associative Memories for Translation , 1997, IWANN.
[32] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[33] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[34] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[35] Marc'Aurelio Ranzato,et al. Learning Longer Memory in Recurrent Neural Networks , 2014, ICLR.
[36] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[37] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[38] Phil Blunsom,et al. Recurrent Continuous Translation Models , 2013, EMNLP.
[39] Xinyun Chen. Under Review as a Conference Paper at Iclr 2017 Delving into Transferable Adversarial Ex- Amples and Black-box Attacks , 2016 .
[40] Holger Schwenk,et al. Continuous space language models , 2007, Comput. Speech Lang..
[41] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.