MMGAN: Generative Adversarial Networks for Multi-Modal Distributions

Over the past years, Generative Adversarial Networks (GANs) have shown a remarkable generation performance especially in image synthesis. Unfortunately, they are also known for having an unstable training process and might loose parts of the data distribution for heterogeneous input data. In this paper, we propose a novel GAN extension for multi-modal distribution learning (MMGAN). In our approach, we model the latent space as a Gaussian mixture model with a number of clusters referring to the number of disconnected data manifolds in the observation space, and include a clustering network, which relates each data manifold to one Gaussian cluster. Thus, the training gets more stable. Moreover, MMGAN allows for clustering real data according to the learned data manifold in the latent space. By a series of benchmark experiments, we illustrate that MMGAN outperforms competitive state-of-the-art models in terms of clustering performance.

[1]  Léon Bottou,et al.  Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.

[2]  G. Lewicki,et al.  Approximation by Superpositions of a Sigmoidal Function , 2003 .

[3]  Avrim Blum,et al.  Foundations of Data Science , 2020 .

[4]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[5]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[6]  Navdeep Jaitly,et al.  Adversarial Autoencoders , 2015, ArXiv.

[7]  Léon Bottou,et al.  Wasserstein Generative Adversarial Networks , 2017, ICML.

[8]  John B. Shoven,et al.  I , Edinburgh Medical and Surgical Journal.

[9]  Pieter Abbeel,et al.  InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.

[10]  Murray Shanahan,et al.  Deep Unsupervised Clustering with Gaussian Mixture Variational Autoencoders , 2016, ArXiv.

[11]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[12]  Sreeram Kannan,et al.  ClusterGAN : Latent Space Clustering in Generative Adversarial Networks , 2018, AAAI.

[13]  Ivan Oseledets,et al.  Universality Theorems for Generative Models , 2019, ArXiv.

[14]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[15]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[16]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[17]  Alexia Jolicoeur-Martineau,et al.  The relativistic discriminator: a key element missing from standard GAN , 2018, ICLR.

[18]  Maneesh Kumar Singh,et al.  Disconnected Manifold Learning for Generative Adversarial Networks , 2018, NeurIPS.

[19]  Daphna Weinshall,et al.  Gaussian Mixture Generative Adversarial Networks for Diverse Datasets, and the Unsupervised Clustering of Images , 2018, ArXiv.

[20]  P ? ? ? ? ? ? ? % ? ? ? ? , 1991 .

[22]  Ravi Kiran Sarvadevabhatla,et al.  DeLiGAN: Generative Adversarial Networks for Diverse and Limited Data , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  J. Jost Riemannian geometry and geometric analysis , 1995 .

[24]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[25]  Tsuyoshi Murata,et al.  {m , 1934, ACML.

[26]  J. Zico Kolter,et al.  Gradient descent GAN optimization is locally stable , 2017, NIPS.