Clustering With Orthogonal AutoEncoder

Recently, clustering algorithms based on deep AutoEncoder attract lots of attention due to their excellent clustering performance. On the other hand, the success of PCA-Kmeans and spectral clustering corroborates that the orthogonality of embedding is beneficial to increase the clustering accuracy. In this paper, we propose a novel dimensional reduction model, called Orthogonal AutoEncoder (OAE), which encourages the orthogonality of the learned embedding. Furthermore, we propose a joint deep Clustering framework based on Orthogonal AutoEncoder (COAE), and this new framework is capable of extracting the latent embedding and predicting the clustering assignment simultaneously. The COAE stacks a fully connected clustering layer on top of the OAE, where the activation function of the clustering layer is the multinomial logistic regression function. The loss function of the COAE contains two terms: the reconstruction loss and the clustering-oriented loss. The first one is a data-dependent term in order to prevent overfitting. The other one is the cross entropy between the predicted assignment and the auxiliary target distribution. The network parameters of the COAE can be effectively updated by the mini-batch stochastic gradient descent algorithm and the back-propagation approach. The experiments on benchmark datasets empirically demonstrate that the COAE can achieve superior or competitive clustering performance as state-of-the-art deep clustering frameworks. The implementation of our algorithm is available at https://github.com/WangDavey/COAE

[1]  Shuigeng Zhou,et al.  DeepCluster: A General Clustering Framework Based on Deep Learning , 2017, ECML/PKDD.

[2]  Tong Zhang,et al.  Deep Subspace Clustering Networks , 2017, NIPS.

[3]  Yi Yang,et al.  Image Clustering Using Local Discriminant Models and Global Integration , 2010, IEEE Transactions on Image Processing.

[4]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[5]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.

[6]  Vladlen Koltun,et al.  Deep Continuous Clustering , 2018, ArXiv.

[7]  Marcin Kurdziel,et al.  Encouraging orthogonality between weight vectors in pretrained deep neural networks , 2016, Neurocomputing.

[8]  Martin Ester,et al.  Frequent term-based text clustering , 2002, KDD.

[9]  Cheng Deng,et al.  Deep Clustering via Joint Convolutional Autoencoder Embedding and Relative Entropy Minimization , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[10]  Guillermo Sapiro,et al.  OLE: Orthogonal Low-rank Embedding, A Plug and Play Geometric Loss for Deep Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[11]  Mathieu Cliche,et al.  BB_twtr at SemEval-2017 Task 4: Twitter Sentiment Analysis with CNNs and LSTMs , 2017, *SEMEVAL.

[12]  Wen Gao,et al.  Manifold-Manifold Distance with application to face recognition based on image set , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  En Zhu,et al.  Deep Clustering with Convolutional Autoencoders , 2017, ICONIP.

[14]  Hujun Bao,et al.  Sparse concept coding for visual analysis , 2011, CVPR 2011.

[15]  Erik Cambria,et al.  Recent Trends in Deep Learning Based Natural Language Processing , 2017, IEEE Comput. Intell. Mag..

[16]  Nanning Zheng,et al.  Entropy and orthogonality based deep discriminative feature learning for object recognition , 2018, Pattern Recognit..

[17]  Francesco Cricri,et al.  Clustering and Unsupervised Anomaly Detection with l2 Normalized Deep Auto-Encoder Representations , 2018, 2018 International Joint Conference on Neural Networks (IJCNN).

[18]  Ali Farhadi,et al.  Unsupervised Deep Embedding for Clustering Analysis , 2015, ICML.

[19]  Bo Yang,et al.  Towards K-means-friendly Spaces: Simultaneous Deep Learning and Clustering , 2016, ICML.

[20]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[21]  Wei-Yun Yau,et al.  Deep Subspace Clustering with Sparsity Prior , 2016, IJCAI.

[22]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[23]  Takeo Kanade,et al.  Discriminative cluster analysis , 2006, ICML.

[24]  D.M. Mount,et al.  An Efficient k-Means Clustering Algorithm: Analysis and Implementation , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[25]  Deng Cai,et al.  Gaussian Mixture Model with Local Consistency , 2010, AAAI.

[26]  Bo Yang,et al.  Learning From Hidden Traits: Joint Factor Analysis and Latent Clustering , 2016, IEEE Transactions on Signal Processing.

[27]  Basura Fernando,et al.  Generalized BackPropagation, Étude De Cas: Orthogonality , 2016, ArXiv.

[28]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  Feng Liu,et al.  Auto-encoder Based Data Clustering , 2013, CIARP.

[30]  Michael I. Jordan,et al.  On Spectral Clustering: Analysis and an algorithm , 2001, NIPS.

[31]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[32]  Jianping Yin,et al.  Improved Deep Embedded Clustering with Local Structure Preservation , 2017, IJCAI.

[33]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[34]  Dhruv Batra,et al.  Joint Unsupervised Learning of Deep Representations and Image Clusters , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  Bo Zhang,et al.  Discriminatively Boosted Image Clustering with Fully Convolutional Auto-Encoders , 2017, Pattern Recognit..

[36]  S. P. Lloyd,et al.  Least squares quantization in PCM , 1982, IEEE Trans. Inf. Theory.

[37]  Surya Ganguli,et al.  Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.

[38]  Qiang Liu,et al.  A Survey of Clustering With Deep Learning: From the Perspective of Network Architecture , 2018, IEEE Access.

[39]  Huachun Tan,et al.  Variational Deep Embedding: An Unsupervised and Generative Approach to Clustering , 2016, IJCAI.

[40]  Vladlen Koltun,et al.  Robust continuous clustering , 2017, Proceedings of the National Academy of Sciences.

[41]  Ronen Basri,et al.  SpectralNet: Spectral Clustering using Deep Neural Networks , 2018, ICLR.

[42]  Zhangyang Wang,et al.  Can We Gain More from Orthogonality Regularizations in Training Deep Networks? , 2018, NeurIPS.

[43]  Chris H. Q. Ding,et al.  K-means clustering via principal component analysis , 2004, ICML.

[44]  Shuigeng Zhou,et al.  Exploiting topic modeling to boost metagenomic reads binning , 2015, BMC Bioinformatics.

[45]  H. Kuhn The Hungarian method for the assignment problem , 1955 .

[46]  René Vidal,et al.  Latent Space Sparse Subspace Clustering , 2013, 2013 IEEE International Conference on Computer Vision.

[47]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[48]  Ali Gholipour,et al.  Semi-Supervised Learning With Deep Embedded Clustering for Image Classification and Segmentation , 2019, IEEE Access.

[49]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[50]  Jiri Matas,et al.  All you need is a good init , 2015, ICLR.

[51]  Christos Boutsidis,et al.  Randomized Dimensionality Reduction for $k$ -Means Clustering , 2011, IEEE Transactions on Information Theory.

[52]  Amar Mitiche,et al.  Deep Clustering: On the Link Between Discriminative Models and K-Means , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.