Sequence generative adversarial nets with a conditional discriminator

Abstract The success of Generative Adversarial Networks (GANs) in image generation attracts researchers to design sequence GANs in text generation. However, the discriminators of those sequence GANs usually provide only one signal per sequence, which can not reflect detailed information, e.g. whether a token is appropriate in a sequence. In addition, maximum likelihood pre-training is typically used in those models, which is time-consuming and obscures the effects of adversarial training. To cope with these problems, we propose a new sequence GAN that consists of a conditional discriminator and a discriminator-augmented generator. The conditional discriminator provides a sequence with token-level signals. The generator is designed to approximate a discriminator-augmented distribution, which avoids pre-training. Experiments show that the conditional discriminator provides more informative guidance, and our model outperforms existing models according to metrics involving both sampling quality and sampling diversity.

[1]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[2]  Alan Ritter,et al.  Adversarial Learning for Neural Dialogue Generation , 2017, EMNLP.

[3]  Vaibhava Goel,et al.  Self-Critical Sequence Training for Image Captioning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Richard Socher,et al.  Pointer Sentinel Mixture Models , 2016, ICLR.

[5]  Zhi Chen,et al.  Adversarial Feature Matching for Text Generation , 2017, ICML.

[6]  Jonathan Berant,et al.  Evaluating Text GANs as Language Models , 2018, NAACL.

[7]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[8]  Kevin Lin,et al.  Adversarial Ranking for Language Generation , 2017, NIPS.

[9]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[10]  Andrew M. Dai,et al.  MaskGAN: Better Text Generation via Filling in the ______ , 2018, ICLR.

[11]  Nina Narodytska,et al.  RelGAN: Relational Generative Adversarial Networks for Text Generation , 2019, ICLR.

[12]  Sepp Hochreiter,et al.  GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.

[13]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[14]  Sandeep Subramanian,et al.  Adversarial Generation of Natural Language , 2017, Rep4NLP@ACL.

[15]  Ben Poole,et al.  Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.

[16]  Xu Sun,et al.  Diversity-Promoting GAN: A Cross-Entropy Based Generative Adversarial Network for Diversified Text Generation , 2018, EMNLP.

[17]  Zhe Gan,et al.  Adversarial Text Generation via Feature-Mover's Distance , 2018, NeurIPS.

[18]  Jun Zhao,et al.  Recurrent Convolutional Neural Networks for Text Classification , 2015, AAAI.

[19]  Jason Weston,et al.  Curriculum learning , 2009, ICML '09.

[20]  Samy Bengio,et al.  Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.

[21]  Yong Yu,et al.  Long Text Generation via Adversarial Training with Leaked Information , 2017, AAAI.

[22]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[23]  Lantao Yu,et al.  SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.