Can Unconditional Language Models Recover Arbitrary Sentences?
暂无分享,去创建一个
Samuel R. Bowman | Kyunghyun Cho | Nishant Subramani | Sam Bowman | Kyunghyun Cho | Nishant Subramani
[1] Eric Jones,et al. SciPy: Open Source Scientific Tools for Python , 2001 .
[2] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[3] Xu Tan,et al. MASS: Masked Sequence to Sequence Pre-training for Language Generation , 2019, ICML.
[4] David J. Field,et al. Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.
[5] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[6] Yoshua Bengio,et al. On Using Monolingual Corpora in Neural Machine Translation , 2015, ArXiv.
[7] Deniz Yuret,et al. Transfer Learning for Low-Resource Neural Machine Translation , 2016, EMNLP.
[8] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[9] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[10] Xiaodong Liu,et al. Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.
[11] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[12] Philipp Koehn,et al. Six Challenges for Neural Machine Translation , 2017, NMT@ACL.
[13] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[14] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[15] David Lopez-Paz,et al. Optimizing the Latent Space of Generative Networks , 2017, ICML.
[16] Steven Bird,et al. NLTK: The Natural Language Toolkit , 2002, ACL.
[17] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[18] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[19] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[20] Sergey Edunov,et al. Pre-trained language model representations for language generation , 2019, NAACL.
[21] Stephen J. Wright,et al. Numerical Optimization , 2018, Fundamental Statistical Inference.
[22] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[23] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[24] Alex Graves,et al. Sequence Transduction with Recurrent Neural Networks , 2012, ArXiv.
[25] Samy Bengio,et al. Generating Sentences from a Continuous Space , 2015, CoNLL.
[26] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[27] Jianfeng Gao,et al. A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.
[28] Jason Yosinski,et al. Measuring the Intrinsic Dimension of Objective Landscapes , 2018, ICLR.
[29] Yoshua Bengio,et al. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches , 2014, SSST@EMNLP.
[30] D K Smith,et al. Numerical Optimization , 2001, J. Oper. Res. Soc..
[31] Pramod Viswanath,et al. Representing Sentences as Low-Rank Subspaces , 2017, ACL.
[32] Chandra Bhagavatula,et al. Semi-supervised sequence tagging with bidirectional language models , 2017, ACL.
[33] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[34] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[35] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[36] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[37] Adam Coates,et al. Cold Fusion: Training Seq2Seq Models Together with Language Models , 2017, INTERSPEECH.
[38] Yoshua Bengio,et al. Context-dependent word representation for neural machine translation , 2016, Comput. Speech Lang..
[39] Lior Wolf,et al. Using the Output Embedding to Improve Language Models , 2016, EACL.
[40] Yun Chen,et al. A Stable and Effective Learning Strategy for Trainable Greedy Decoding , 2018, EMNLP.
[41] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[42] Kyunghyun Cho,et al. Noisy Parallel Approximate Decoding for Conditional Recurrent Language Model , 2016, ArXiv.