Locality Sensitive Discriminative Unsupervised Dimensionality Reduction

Graph-based embedding methods receive much attention due to the use of graph and manifold information. However, conventional graph-based embedding methods may not always be effective if the data have high dimensions and have complex distributions. First, the similarity matrix only considers local distance measurement in the original space, which cannot reflect a wide variety of data structures. Second, separation of graph construction and dimensionality reduction leads to the similarity matrix not being fully relied on because the original data usually contain lots of noise samples and features. In this paper, we address these problems by constructing two adjacency graphs to stand for the original structure featuring similarity and diversity of the data, and then impose a rank constraint on the corresponding Laplacian matrix to build a novel adaptive graph learning method, namely locality sensitive discriminative unsupervised dimensionality reduction (LSDUDR). As a result, the learned graph shows a clear block diagonal structure so that the clustering structure of data can be preserved. Experimental results on synthetic datasets and real-world benchmark data sets demonstrate the effectiveness of our approach.

[1]  Chris H. Q. Ding,et al.  Non-negative Laplacian Embedding , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[2]  K. Fan On a Theorem of Weyl Concerning Eigenvalues of Linear Transformations: II. , 1949, Proceedings of the National Academy of Sciences of the United States of America.

[3]  B. Mohar THE LAPLACIAN SPECTRUM OF GRAPHS y , 1991 .

[4]  Xuelong Li,et al.  Initialization Independent Clustering With Actively Self-Training Method , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[5]  Qiang Cheng,et al.  The Fisher-Markov Selector: Fast Selecting Maximally Separable Feature Subset for Multiclass Classification with Applications to High-Dimensional Data , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[6]  Paul Geladi,et al.  Principal Component Analysis , 1987, Comprehensive Chemometrics.

[7]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[8]  Feiping Nie,et al.  The Constrained Laplacian Rank Algorithm for Graph-Based Clustering , 2016, AAAI.

[9]  K. Fan On a Theorem of Weyl Concerning Eigenvalues of Linear Transformations I. , 1949, Proceedings of the National Academy of Sciences of the United States of America.

[10]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .

[11]  Wenjun Zhou,et al.  Efficient locality weighted sparse representation for graph-based learning , 2017, Knowl. Based Syst..

[12]  Feiping Nie,et al.  Unsupervised and semi-supervised learning via ℓ1-norm graph , 2011, 2011 International Conference on Computer Vision.

[13]  Rinkle Rani,et al.  Local graph based correlation clustering , 2017, Knowl. Based Syst..

[14]  Xinbo Gao,et al.  A novel semi-supervised learning for face recognition , 2015, Neurocomputing.

[15]  Huan Liu,et al.  Feature Selection for Clustering , 2000, Encyclopedia of Database Systems.

[16]  Brian C. Lovell,et al.  Graph embedding discriminant analysis on Grassmannian manifolds for improved image set matching , 2011, CVPR 2011.

[17]  Qi Fan,et al.  Multiple empirical kernel learning with locality preserving constraint , 2016, Knowl. Based Syst..

[18]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[19]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[20]  Moshe Ben-Bassat,et al.  35 Use of distance measures, information measures and error bounds in feature evaluation , 1982, Classification, Pattern Recognition and Reduction of Dimensionality.

[21]  Liang Du,et al.  Unsupervised Feature Selection with Adaptive Structure Learning , 2015, KDD.

[22]  R. Fisher THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMS , 1936 .

[23]  Rui Xu,et al.  Survey of clustering algorithms , 2005, IEEE Transactions on Neural Networks.

[24]  Feiping Nie,et al.  Cauchy Graph Embedding , 2011, ICML.

[25]  L. N. Kanal,et al.  Handbook of Statistics, Vol. 2. Classification, Pattern Recognition and Reduction of Dimensionality. , 1985 .

[26]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[27]  James C. Bezdek,et al.  Convergence of Alternating Optimization , 2003, Neural Parallel Sci. Comput..

[28]  Weifu Chen,et al.  Spectral clustering: A semi-supervised approach , 2012, Neurocomputing.

[29]  Feiping Nie,et al.  Clustering and projected clustering with adaptive neighbors , 2014, KDD.

[30]  Pietro Perona,et al.  Self-Tuning Spectral Clustering , 2004, NIPS.

[31]  Andrew B. Kahng,et al.  New spectral methods for ratio cut partitioning and clustering , 1991, IEEE Trans. Comput. Aided Des. Integr. Circuits Syst..

[32]  Kenneth M. Hall An r-Dimensional Quadratic Placement Algorithm , 1970 .

[33]  Hailin Zhang,et al.  Joint Global and Local Structure Discriminant Analysis , 2013, IEEE Trans. Inf. Forensics Secur..

[34]  Zhong Jin,et al.  Face recognition using discriminant sparsity neighborhood preserving embedding , 2012, Knowl. Based Syst..

[35]  Bin Luo,et al.  Similarity Learning of Manifold Data , 2015, IEEE Transactions on Cybernetics.

[36]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[37]  Feiping Nie,et al.  Neighborhood MinMax Projections , 2007, IJCAI.

[38]  Feiping Nie,et al.  Semi-supervised orthogonal discriminant analysis via label propagation , 2009, Pattern Recognit..

[39]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.