暂无分享,去创建一个
[1] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[2] Yonghui Wu,et al. Exploring the Limits of Language Modeling , 2016, ArXiv.
[3] Yee Whye Teh,et al. A fast and simple algorithm for training neural probabilistic language models , 2012, ICML.
[4] Aapo Hyvärinen,et al. Noise-Contrastive Estimation of Unnormalized Statistical Models, with Applications to Natural Image Statistics , 2012, J. Mach. Learn. Res..
[5] Adam L. Berger,et al. A Maximum Entropy Approach to Natural Language Processing , 1996, CL.
[6] Andrew McCallum,et al. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data , 2001, ICML.
[7] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[8] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[9] Yoshua Bengio,et al. Adaptive Importance Sampling to Accelerate Training of a Neural Probabilistic Language Model , 2008, IEEE Transactions on Neural Networks.
[10] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[11] T. Ferguson. A Course in Large Sample Theory , 1996 .
[12] Omer Levy,et al. Neural Word Embedding as Implicit Matrix Factorization , 2014, NIPS.