Reducing infrequent-token perplexity via variational corpora
暂无分享,去创建一个
[1] George A. Miller,et al. WordNet: A Lexical Database for English , 1995, HLT.
[2] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[3] Hinrich Schütze,et al. Book Reviews: Foundations of Statistical Natural Language Processing , 1999, CL.
[4] Hayato Kobayashi. Perplexity on Reduced Corpora , 2014, ACL.
[5] Min Zhang,et al. Improving Language Model Size Reduction using Better Pruning Criteria , 2002, ACL.
[6] Slava M. Katz,et al. Estimation of probabilities from sparse data for the language model component of a speech recognizer , 1987, IEEE Trans. Acoust. Speech Signal Process..
[7] Andruid Kerne,et al. Everyday ideation: all of my ideas are on pinterest , 2014, CHI.
[8] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[9] Geoffrey E. Hinton,et al. Generating Text with Recurrent Neural Networks , 2011, ICML.
[10] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.
[11] Yee Whye Teh,et al. A Bayesian Interpretation of Interpolated Kneser-Ney , 2006 .
[12] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[13] Ming Zhou,et al. A Recursive Recurrent Neural Network for Statistical Machine Translation , 2014, ACL.
[14] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[15] Phil Blunsom,et al. Compositional Morphology for Word Representations and Language Modelling , 2014, ICML.
[16] Jianfeng Gao,et al. Language model size reduction by pruning and clustering , 2000, INTERSPEECH.
[17] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[18] Meeyoung Cha,et al. Social bootstrapping: how pinterest and last.fm social communities benefit by borrowing links from facebook , 2014, WWW.
[19] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[20] Kenneth Ward Church,et al. Compressing Trigram Language Models With Golomb Coding , 2007, EMNLP.
[21] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[22] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[23] G. Zipf. The Psycho-Biology Of Language: AN INTRODUCTION TO DYNAMIC PHILOLOGY , 1999 .