TextGAIL: Generative Adversarial Imitation Learning for Text Generation
暂无分享,去创建一个
Qingyang Wu | Zhou Yu | Lei Li | Lei Li | Zhou Yu | Qingyang Wu
[1] Anind K. Dey,et al. Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.
[2] Jonathan Berant,et al. Evaluating Text GANs as Language Models , 2018, NAACL.
[3] Joelle Pineau,et al. A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues , 2016, AAAI.
[4] Nathanael Chambers,et al. A Corpus and Evaluation Framework for Deeper Understanding of Commonsense Stories , 2016, ArXiv.
[5] Minlie Huang,et al. ARAML: A Stable Adversarial Training Framework for Text Generation , 2019, EMNLP.
[6] Pei Zhou,et al. CommonGen: A Constrained Text Generation Dataset Towards Generative Commonsense Reasoning , 2019, ArXiv.
[7] Min Lin,et al. Softmax GAN , 2017, ArXiv.
[8] Quoc V. Le,et al. Towards a Human-like Open-Domain Chatbot , 2020, ArXiv.
[9] Yong Yu,et al. Long Text Generation via Adversarial Training with Leaked Information , 2017, AAAI.
[10] Rouhollah Rahmatizadeh,et al. Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-to-End Learning from Demonstration , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[11] Jianfeng Gao,et al. A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.
[12] Kevin Lin,et al. Adversarial Ranking for Language Generation , 2017, NIPS.
[13] Jason Weston,et al. Neural Text Generation with Unlikelihood Training , 2019, ICLR.
[14] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[15] Quoc V. Le,et al. A Simple Method for Commonsense Reasoning , 2018, ArXiv.
[16] J. Andrew Bagnell,et al. Efficient Reductions for Imitation Learning , 2010, AISTATS.
[17] Bowen Zhou,et al. Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond , 2016, CoNLL.
[18] Ali Farhadi,et al. Defending Against Neural Fake News , 2019, NeurIPS.
[19] Nina Narodytska,et al. RelGAN: Relational Generative Adversarial Networks for Text Generation , 2019, ICLR.
[20] Alexia Jolicoeur-Martineau,et al. The relativistic discriminator: a key element missing from standard GAN , 2018, ICLR.
[21] Ke Xu,et al. Self-Adversarial Learning with Comparative Discrimination for Text Generation , 2020, ICLR.
[22] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[23] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[24] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[25] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[26] Matt J. Kusner,et al. GANS for Sequences of Discrete Elements with the Gumbel-softmax Distribution , 2016, ArXiv.
[27] Yoshua Bengio,et al. Maximum-Likelihood Augmented Discrete Generative Adversarial Networks , 2017, ArXiv.
[28] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[29] Stanislau Semeniuta,et al. On Accurate Evaluation of GANs for Language Generation , 2018, ArXiv.
[30] Mary Williamson,et al. Recipes for Building an Open-Domain Chatbot , 2020, EACL.
[31] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[32] Richard Tanburn,et al. Making Efficient Use of Demonstrations to Solve Hard Exploration Problems , 2019, ICLR.
[33] Anca D. Dragan,et al. SQIL: Imitation Learning via Reinforcement Learning with Sparse Rewards , 2019, ICLR.
[34] Yejin Choi,et al. The Curious Case of Neural Text Degeneration , 2019, ICLR.
[35] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[36] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[37] Joelle Pineau,et al. Language GANs Falling Short , 2018, ICLR.
[38] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[39] Marc'Aurelio Ranzato,et al. Sequence Level Training with Recurrent Neural Networks , 2015, ICLR.
[40] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[41] Alexander M. Rush,et al. Sequence-to-Sequence Learning as Beam-Search Optimization , 2016, EMNLP.
[42] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.