GeDi: Generative Discriminator Guided Sequence Generation
暂无分享,去创建一个
Richard Socher | Nazneen Fatema Rajani | Bryan McCann | Shafiq Joty | Ben Krause | Nitish Shirish Keskar | Akhilesh Deepak Gotmare
[1] Noah A. Smith,et al. Contrastive Estimation: Training Log-Linear Models on Unlabeled Data , 2005, ACL.
[2] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[3] Lav R. Varshney,et al. CTRL: A Conditional Transformer Language Model for Controllable Generation , 2019, ArXiv.
[4] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[5] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[6] Eric P. Xing,et al. Toward Controlled Generation of Text , 2017, ICML.
[7] Wang Ling,et al. Generative and Discriminative Text Classification with Recurrent Neural Networks , 2017, ArXiv.
[8] John Scott Bridle,et al. Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition , 1989, NATO Neurocomputing.
[9] Ilya Sutskever,et al. Learning to Generate Reviews and Discovering Sentiment , 2017, ArXiv.
[10] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[11] Andrew M. Dai,et al. MaskGAN: Better Text Generation via Filling in the ______ , 2018, ICLR.
[12] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[13] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[14] Vasant Honavar,et al. Discriminatively trained Markov model for sequence classification , 2005, Fifth IEEE International Conference on Data Mining (ICDM'05).
[15] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[16] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[17] Jason Yosinski,et al. Plug and Play Language Models: A Simple Approach to Controlled Text Generation , 2020, ICLR.
[18] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[19] Xing Shi,et al. Hafez: an Interactive Poetry Generation System , 2017, ACL.
[20] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[21] Yoav Goldberg,et al. Controlling Linguistic Style Aspects in Neural Language Generation , 2017, ArXiv.
[22] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[23] Yejin Choi,et al. The Curious Case of Neural Text Degeneration , 2019, ICLR.
[24] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[25] Alec Radford,et al. Fine-Tuning Language Models from Human Preferences , 2019, ArXiv.
[26] Hao Liu,et al. Hybrid Discriminative-Generative Training via Contrastive Learning , 2020, ArXiv.
[27] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[28] Richard Socher,et al. Explain Yourself! Leveraging Language Models for Commonsense Reasoning , 2019, ACL.
[29] Tom Minka,et al. Principled Hybrids of Generative and Discriminative Models , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).
[30] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[31] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[32] Yee Whye Teh,et al. A fast and simple algorithm for training neural probabilistic language models , 2012, ICML.
[33] Mirella Lapata,et al. Text Summarization with Pretrained Encoders , 2019, EMNLP.
[34] Yejin Choi,et al. Learning to Write with Cooperative Discriminators , 2018, ACL.
[35] Michael I. Jordan,et al. On Discriminative vs. Generative Classifiers: A comparison of logistic regression and naive Bayes , 2001, NIPS.
[36] Dragomir R. Radev,et al. Multi-News: A Large-Scale Multi-Document Summarization Dataset and Abstractive Hierarchical Model , 2019, ACL.
[37] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[38] Thomas Wolf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[39] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .