Clustering Based on Supervised Learning of Exemplar Discriminative Information

In machine learning and data mining applications, clustering is a critical task for knowledge discovery that attract attentions from large quantities of researchers. Generally, with the help of label information, supervised learning methods have more flexible structure and better result than unsupervised learning. However, supervised learning is infeasible for clustering task. In this paper, to fill the gap between clustering and supervised learning, the proposed clustering methods introduce the exemplars discriminative information into a supervised learning. To build the effective objective function, a strategy for reducing intracluster distance and increasing intercluster distance is introduced to form a unified optimization objective function. With initially setting the clustering centers, the data that near the centers are selected as exemplars to indicate the ground truth of different classes. Discriminative learning is then introduced to learn the partition hyperplane and classify all the data into different classes. New clustering centers are calculated for selecting new exemplars alternately. Using the proposed algorithms, the unsupervised $K$ -means clustering problem is effectively solved from the perceptive of optimization. Feature mapping is also introduced to improve the performance by reducing the intercluster distance. A novel framework for exploring discriminative information from unsupervised data is provided. The proposed algorithms outperform the state-of-the-art approaches on a wide range of benchmark datasets in terms of accuracy.

[1]  Jonathan J. Hull,et al.  A Database for Handwritten Text Recognition Research , 1994, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Guang-Bin Huang,et al.  Extreme learning machine: a new learning scheme of feedforward neural networks , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[3]  Laurent Navarro,et al.  Anomaly Detection with the Voronoi Diagram Evolutionary Algorithm , 2016, PPSN.

[4]  Ognjen Arandjelovic Discriminative k-means clustering , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).

[5]  D. Tran,et al.  Fuzzy entropy clustering , 2000, Ninth IEEE International Conference on Fuzzy Systems. FUZZ- IEEE 2000 (Cat. No.00CH37063).

[6]  Kurt Hornik,et al.  Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.

[7]  Dilip B. Kotak,et al.  GRIDBSCAN: GRId Density-Based Spatial Clustering of Applications with Noise , 2006, 2006 IEEE International Conference on Systems, Man and Cybernetics.

[8]  Teuvo Kohonen,et al.  Self-organized formation of topologically correct feature maps , 2004, Biological Cybernetics.

[9]  Delbert Dueck,et al.  Clustering by Passing Messages Between Data Points , 2007, Science.

[10]  Hui Xiong,et al.  K-means clustering versus validation measures: a data distribution perspective , 2006, KDD '06.

[11]  Cheng Wu,et al.  Discriminative clustering via extreme learning machine , 2015, Neural Networks.

[12]  Andy Harter,et al.  Parameterisation of a stochastic model for human face identification , 1994, Proceedings of 1994 IEEE Workshop on Applications of Computer Vision.

[13]  Cheng Wu,et al.  Semi-Supervised and Unsupervised Extreme Learning Machines , 2014, IEEE Transactions on Cybernetics.

[14]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[15]  Shi-Yuan Han,et al.  K-medoids method based on divergence for uncertain data clustering , 2016, 2016 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[16]  C. L. Philip Chen,et al.  A Collaborative Fuzzy Clustering Algorithm in Distributed Network Environments , 2014, IEEE Transactions on Fuzzy Systems.

[17]  Taghi M. Khoshgoftaar,et al.  Software Quality Analysis of Unlabeled Program Modules With Semisupervised Clustering , 2007, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[18]  Charu C. Aggarwal,et al.  A Survey of Text Clustering Algorithms , 2012, Mining Text Data.

[19]  Zhiping Lin,et al.  Extreme Learning Machine for Clustering , 2015 .

[20]  Christopher M. Bishop,et al.  Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .

[21]  Adrian E. Raftery,et al.  Model-Based Clustering, Discriminant Analysis, and Density Estimation , 2002 .

[22]  C. L. Philip Chen,et al.  A rapid learning and dynamic stepwise updating algorithm for flat neural networks and the application to time-series prediction , 1999, IEEE Trans. Syst. Man Cybern. Part B.

[23]  Yuan Zhang,et al.  Fuzzy clustering with the entropy of attribute weights , 2016, Neurocomputing.

[24]  Chunxia Zhang,et al.  Generalized extreme learning machine autoencoder and a new deep neural network , 2017, Neurocomputing.

[25]  Nikolaos G. Bourbakis,et al.  Graph-Based Methods for Natural Language Processing and Understanding—A Survey and Analysis , 2014, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[26]  Guang-Bin Huang,et al.  An Insight into Extreme Learning Machines: Random Neurons, Random Features and Kernels , 2014, Cognitive Computation.

[27]  Chen Zhang,et al.  Extreme Maximum Margin Clustering , 2013, IEICE Trans. Inf. Syst..

[28]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[29]  Fuzhen Zhuang,et al.  Clustering in extreme learning machine feature space , 2014, Neurocomputing.

[30]  Gang Niu,et al.  Bayesian Maximum Margin Clustering , 2010, 2010 IEEE International Conference on Data Mining.

[31]  Jieping Ye,et al.  Discriminative K-means for Clustering , 2007, NIPS.

[32]  Chris H. Q. Ding,et al.  K-means clustering via principal component analysis , 2004, ICML.

[33]  Victor C. M. Leung,et al.  Extreme Learning Machines [Trends & Controversies] , 2013, IEEE Intelligent Systems.

[34]  Ivor W. Tsang,et al.  Maximum Margin Clustering Made Practical , 2007, IEEE Transactions on Neural Networks.

[35]  Jürgen Schmidhuber,et al.  Training Very Deep Networks , 2015, NIPS.

[36]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[37]  Gábor J. Székely,et al.  Hierarchical Clustering via Joint Between-Within Distances: Extending Ward's Minimum Variance Method , 2005, J. Classif..

[38]  Jiong Yang,et al.  STING: A Statistical Information Grid Approach to Spatial Data Mining , 1997, VLDB.

[39]  Dale Schuurmans,et al.  Maximum Margin Clustering , 2004, NIPS.

[40]  Anna Fabijańska Normalized cuts and watersheds for image segmentation , 2012 .

[41]  Nan Zhang,et al.  Unsupervised and semi-supervised extreme learning machine with wavelet kernel for high dimensional data , 2016, Memetic Computing.

[42]  C. L. Philip Chen,et al.  Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[43]  C. L. Philip Chen,et al.  A Cooperative Learning-Based Clustering Approach to Lip Segmentation Without Knowing Segment Number , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[44]  Zhongzhi Shi,et al.  Unsupervised extreme learning machine with representational features , 2015, International Journal of Machine Learning and Cybernetics.

[45]  Feiping Nie,et al.  Discriminative Embedded Clustering: A Framework for Grouping High-Dimensional Data , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[46]  Yi-Shin Chen,et al.  SociRank: Identifying and Ranking Prevalent News Topics Using Social Media Factors , 2017, IEEE Transactions on Systems, Man, and Cybernetics: Systems.