Neural Topic Modeling with Cycle-Consistent Adversarial Training

Advances on deep generative models have attracted significant research interest in neural topic modeling. The recently proposed Adversarial-neural Topic Model models topics with an adversarially trained generator network and employs Dirichlet prior to capture the semantic patterns in latent topics. It is effective in discovering coherent topics but unable to infer topic distributions for given documents or utilize available document labels. To overcome such limitations, we propose Topic Modeling with Cycle-consistent Adversarial Training (ToMCAT) and its supervised version sToMCAT. ToMCAT employs a generator network to interpret topics and an encoder network to infer document topics. Adversarial training and cycle-consistent constraints are used to encourage the generator and the encoder to produce realistic samples that coordinate with each other. sToMCAT extends ToMCAT by incorporating document labels into the topic modeling process to help discover more coherent topics. The effectiveness of the proposed models is evaluated on unsupervised/supervised topic modeling and text classification. The experimental results show that our models can produce both coherent and informative topics, outperforming a number of competitive baselines.

[1]  Eric P. Xing,et al.  Sparse Additive Generative Models of Text , 2011, ICML.

[2]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[3]  David M. Blei,et al.  Supervised Topic Models , 2007, NIPS.

[4]  Rui Wang,et al.  Open Event Extraction from Online Text using a Generative Adversarial Network , 2019, EMNLP.

[5]  Mark Stevenson,et al.  Evaluating Topic Coherence Using Distributional Semantics , 2013, IWCS.

[6]  Phil Blunsom,et al.  Discovering Discrete Latent Topics with Neural Variational Inference , 2017, ICML.

[7]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[8]  Phil Blunsom,et al.  Neural Variational Inference for Text Processing , 2015, ICML.

[9]  David J. C. MacKay,et al.  Choice of Basis for Laplace Approximation , 1998, Machine Learning.

[10]  Rui Wang,et al.  ATM: Adversarial-neural Topic Model , 2018, Inf. Process. Manag..

[11]  Léon Bottou,et al.  Wasserstein Generative Adversarial Networks , 2017, ICML.

[12]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[13]  Timothy Baldwin,et al.  Automatic Evaluation of Topic Coherence , 2010, NAACL.

[14]  Xuan Zhang,et al.  Event extraction from Twitter using Non-Parametric Bayesian Mixture Model with Word Embeddings , 2017, EACL.

[15]  Deyu Zhou,et al.  Neural Topic Modeling with Bidirectional Adversarial Training , 2020, ACL.

[16]  Hyunsoo Kim,et al.  Learning to Discover Cross-Domain Relations with Generative Adversarial Networks , 2017, ICML.

[17]  Trevor Darrell,et al.  Adversarial Feature Learning , 2016, ICLR.

[18]  Yulan He,et al.  Joint sentiment/topic model for sentiment analysis , 2009, CIKM.

[19]  Aaron C. Courville,et al.  Improved Training of Wasserstein GANs , 2017, NIPS.

[20]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[21]  Noah A. Smith,et al.  Neural Models for Documents with Metadata , 2017, ACL.

[22]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[23]  拓海 杉山,et al.  “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”の学習報告 , 2017 .

[24]  Michael Röder,et al.  Exploring the Space of Topic Coherence Measures , 2015, WSDM.

[25]  Lin-Shan Lee,et al.  Scalable Sentiment for Sequence-to-Sequence Chatbot Response with Performance Analysis , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[26]  Mark Steyvers,et al.  Finding scientific topics , 2004, Proceedings of the National Academy of Sciences of the United States of America.

[27]  Andrew McCallum,et al.  Rethinking LDA: Why Priors Matter , 2009, NIPS.

[28]  Charles A. Sutton,et al.  Autoencoding Variational Inference For Topic Models , 2017, ICLR.

[29]  Andrew L. Maas Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .

[30]  Xiang Zhang,et al.  Character-level Convolutional Networks for Text Classification , 2015, NIPS.

[31]  Gerlof Bouma,et al.  Normalized (pointwise) mutual information in collocation extraction , 2009 .