Discriminative Clustering of High-Dimensional Data Using Generative Modeling

We approach unsupervised clustering from a generative perspective. We hybridize Variational Autoencoder (VAE) and Generative Adversarial Network (GAN) in a novel way to obtain a vigorous clustering model that can effectively be applied to challenging high-dimensional datasets. The powerful inference of the VAE is used along with a categorical discriminator that aims to obtain a cluster assignment of the data, by maximizing the mutual information between the observations and their predicted class distribution. The discriminator is regularized with examples produced by an adversarial generator, whose task is to trick the discriminator into accepting them as real data. We demonstrate that using a shared latent representation greatly helps with discriminative power of our model and leads to a powerful unsupervised clustering model. The method can be applied to raw data in a high-dimensional space. Training can be performed end-to-end from randomly-initialized weights by alternating stochastic gradient descent on the parameters of the model. Experiments on two datasets including the challenging MNIST dataset show that the proposed method performs better than the existing models. Additionally, our method yields an efficient generative model.

[1]  Ali Farhadi,et al.  Unsupervised Deep Embedding for Clustering Analysis , 2015, ICML.

[2]  David J. C. MacKay,et al.  Unsupervised Classifiers, Mutual Information and 'Phantom Targets' , 1991, NIPS.

[3]  Navdeep Jaitly,et al.  Adversarial Autoencoders , 2015, ArXiv.

[4]  Max Welling,et al.  Auto-Encoding Variational Bayes , 2013, ICLR.

[5]  Donald C. Wunsch,et al.  Clustering Data of Mixed Categorical and Numerical Type With Unsupervised Feature Learning , 2015, IEEE Access.

[6]  Murray Shanahan,et al.  Deep Unsupervised Clustering with Gaussian Mixture Variational Autoencoders , 2016, ArXiv.

[7]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[8]  Charu C. Aggarwal,et al.  Data Clustering: Algorithms and Applications , 2014 .

[9]  Brendan J. Frey,et al.  PixelGAN Autoencoders , 2017, NIPS.

[10]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[11]  Pascal Poupart,et al.  Generative mixture of networks , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[12]  Saeid Nahavandi,et al.  Multi-Residual Networks , 2016, ArXiv.

[13]  Jost Tobias Springenberg,et al.  Unsupervised and Semi-supervised Learning with Categorical Generative Adversarial Networks , 2015, ICLR.

[14]  Ali Borji,et al.  A new look at clustering through the lens of deep convolutional neural networks , 2017, ArXiv.

[15]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[16]  Christopher Burgess,et al.  beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.

[17]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[18]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[19]  Andreas Krause,et al.  Discriminative Clustering by Regularized Information Maximization , 2010, NIPS.

[20]  Pieter Abbeel,et al.  InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.

[21]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[22]  E. M. Wright,et al.  Adaptive Control Processes: A Guided Tour , 1961, The Mathematical Gazette.

[23]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.