How to represent a word and predict it, too: Improving tied architectures for language modelling
暂无分享,去创建一个
[1] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[2] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[3] Hermann Ney,et al. LSTM Neural Networks for Language Modeling , 2012, INTERSPEECH.
[4] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[5] Lior Wolf,et al. Using the Output Embedding to Improve Language Models , 2016, EACL.
[6] Sanjiv Kumar,et al. On the Convergence of Adam and Beyond , 2018 .
[7] Hakan Inan,et al. Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling , 2016, ICLR.
[8] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[9] Christopher D. Manning,et al. Better Word Representations with Recursive Neural Networks for Morphology , 2013, CoNLL.
[10] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[11] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[12] Felix Hill,et al. SimLex-999: Evaluating Semantic Models With (Genuine) Similarity Estimation , 2014, CL.
[13] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[14] Lukás Burget,et al. Extensions of recurrent neural network language model , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[15] Elia Bruni,et al. Multimodal Distributional Semantics , 2014, J. Artif. Intell. Res..