Pre-train and Plug-in: Flexible Conditional Text Generation with Variational Auto-Encoders
暂无分享,去创建一个
[1] Ming Li,et al. Generating Thematic Chinese Poetry using Conditional Variational Autoencoders with Hybrid Decoders , 2017, IJCAI.
[2] Luc Van Gool,et al. Wasserstein Divergence for GANs , 2017, ECCV.
[3] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[4] Simon Osindero,et al. Conditional Generative Adversarial Nets , 2014, ArXiv.
[5] Jeff Donahue,et al. Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.
[6] Adam Roberts,et al. Latent Constraints: Learning to Generate Conditionally from Unconditional Generative Models , 2017, ICLR.
[7] Mohit Bansal,et al. Polite Dialogue Generation Without Parallel Data , 2018, TACL.
[8] Eric P. Xing,et al. Toward Controlled Generation of Text , 2017, ICML.
[9] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[10] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[11] Jason Weston,et al. What makes a good conversation? How controllable attributes affect human judgments , 2019, NAACL.
[12] Andriy Mnih,et al. Disentangling by Factorising , 2018, ICML.
[13] Ming Zhou,et al. Question Generation from SQL Queries Improves Neural Semantic Parsing , 2018, EMNLP.
[14] Yann Dauphin,et al. Hierarchical Neural Story Generation , 2018, ACL.
[15] Ke Xu,et al. Self-Adversarial Learning with Comparative Discrimination for Text Generation , 2020, ICLR.
[16] Emilien Dupont,et al. Joint-VAE: Learning Disentangled Joint Continuous and Discrete Representations , 2018, NeurIPS.
[17] Ali Razavi,et al. Generating Diverse High-Fidelity Images with VQ-VAE-2 , 2019, NeurIPS.
[18] Sebastian Ruder,et al. Fine-tuned Language Models for Text Classification , 2018, ArXiv.
[19] Bernhard Schölkopf,et al. Wasserstein Auto-Encoders , 2017, ICLR.
[20] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[21] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[22] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[23] Maxine Eskénazi,et al. Learning Discourse-level Diversity for Neural Dialog Models using Conditional Variational Autoencoders , 2017, ACL.
[24] Lav R. Varshney,et al. CTRL: A Conditional Transformer Language Model for Controllable Generation , 2019, ArXiv.
[25] Roger B. Grosse,et al. Isolating Sources of Disentanglement in Variational Autoencoders , 2018, NeurIPS.
[26] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[27] Yoav Goldberg,et al. Controlling Linguistic Style Aspects in Neural Language Generation , 2017, ArXiv.
[28] Maosong Sun,et al. Chinese Poetry Generation with a Working Memory Model , 2018, IJCAI.
[29] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[30] Svetlana Lazebnik,et al. Diverse and Accurate Image Description Using a Variational Auto-Encoder with an Additive Gaussian Encoding Space , 2017, NIPS.
[31] Jianfeng Gao,et al. A Diversity-Promoting Objective Function for Neural Conversation Models , 2015, NAACL.
[32] Honglak Lee,et al. Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.
[33] Dongyan Zhao,et al. Style Transfer in Text: Exploration and Evaluation , 2017, AAAI.
[34] Regina Barzilay,et al. Style Transfer from Non-Parallel Text by Cross-Alignment , 2017, NIPS.
[35] Graham Neubig,et al. Controlling Output Length in Neural Encoder-Decoders , 2016, EMNLP.
[36] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[37] Samy Bengio,et al. Generating Sentences from a Continuous Space , 2015, CoNLL.
[38] Mona Attariyan,et al. Parameter-Efficient Transfer Learning for NLP , 2019, ICML.
[39] Ke Wang,et al. SentiGAN: Generating Sentimental Texts via Mixture Adversarial Networks , 2018, IJCAI.
[40] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[41] Max Welling,et al. Semi-supervised Learning with Deep Generative Models , 2014, NIPS.