暂无分享,去创建一个
[1] Laure Thompson,et al. The strange geometry of skip-gram with negative sampling , 2017, EMNLP.
[2] Andreas Stolcke,et al. SRILM - an extensible language modeling toolkit , 2002, INTERSPEECH.
[3] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[4] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[5] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[6] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[7] Moustapha Cissé,et al. Efficient softmax approximation for GPUs , 2016, ICML.
[8] Ruslan Salakhutdinov,et al. Breaking the Softmax Bottleneck: A High-Rank RNN Language Model , 2017, ICLR.
[9] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[10] David P. Dobkin,et al. The quickhull algorithm for convex hulls , 1996, TOMS.
[11] Rada Mihalcea,et al. Factors Influencing the Surprising Instability of Word Embeddings , 2018, NAACL.
[12] Pascal Vincent,et al. An Exploration of Softmax Alternatives Belonging to the Spherical Loss Family , 2015, ICLR.
[13] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[14] Richard Socher,et al. Pointer Sentinel Mixture Models , 2016, ICLR.