Linear dimensionality reduction based on Hybrid structure preserving projections

Recent advances have shown the methods based on local structure preserving projections can effectively learn discriminative features. The two attractive approaches for characterizing such data structure are: the classical nearest neighbor strategy for neighborhood structure and the sparse coding algorithm for sparsity structure. Motivated by the intuitive analysis of the relationship between the two structures, in this paper, we take both of them into account and propose two integrated approaches for dimensionality reduction. Concretely, we for achieving improvement directly integrate two available objectives, utilizing neighborhood structure and based on sparsity structure, to construct the combined method, briefly called CSNP. However, such rough strategy often results in its degradation in practice. Instead of the superficial combination, we exploit a hybrid structure by intergrading the two structures and then propose the Sparsity and Neighborhood Preserving Projections, dubbed SNPP, by preserving the hybrid structure into reduced subspace. The resulting optimization problems can be also interpreted as an instance of the general graph embedding framework and can reduce to the generalized eigenvalue decomposition problem. Finally, we conduct extensive experiments on publicly available data sets to verify the efficacy of our algorithms. From the experimental results, we roughly draw the conclusion that neighborhood structure is more important for low-dimensional data while sparsity structure is more useful for high-dimensional data. Construct a hybrid representation to characterize data structure.Perform dimensionality reduction by preserving a hybrid structure.Explicitly discuss the sparsity and the neighborhood depending on datasets.Take multiple structures into account for better algorithm performance.

[1]  Bin Shen,et al.  Learning dictionary on manifolds for image classification , 2013, Pattern Recognit..

[2]  Trevor F. Cox,et al.  Metric multidimensional scaling , 2000 .

[3]  Jiawei Han,et al.  Isometric Projection , 2007, AAAI.

[4]  Stephen Lin,et al.  Graph Embedding and Extensions: A General Framework for Dimensionality Reduction , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  David J. Kriegman,et al.  From Few to Many: Illumination Cone Models for Face Recognition under Variable Lighting and Pose , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[6]  Matti Pietikäinen,et al.  Supervised Locally Linear Embedding , 2003, ICANN.

[7]  Anil K. Jain,et al.  Statistical Pattern Recognition: A Review , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[8]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Shuang Wang,et al.  Fast Fisher Sparsity Preserving Projections , 2012, Neural Computing and Applications.

[10]  Daoqiang Zhang,et al.  Semi-Supervised Dimensionality Reduction ∗ , 2007 .

[11]  Shuicheng Yan,et al.  Neighborhood preserving embedding , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[12]  S. Mallat,et al.  Adaptive greedy approximations , 1997 .

[13]  Masashi Sugiyama,et al.  Dimensionality Reduction of Multimodal Labeled Data by Local Fisher Discriminant Analysis , 2007, J. Mach. Learn. Res..

[14]  Hongtao Lu,et al.  Efficient linear discriminant analysis with locality preserving for face recognition , 2012, Pattern Recognit..

[15]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[16]  Eric O. Postma,et al.  Dimensionality Reduction: A Comparative Review , 2008 .

[17]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[18]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[19]  John Ong,et al.  Compressed Sensing : Theory and Applications in Wireless Neural Recording , 2011 .

[20]  Stéphane Mallat,et al.  A Wavelet Tour of Signal Processing - The Sparse Way, 3rd Edition , 2008 .

[21]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[22]  René Vidal,et al.  Sparse Subspace Clustering: Algorithm, Theory, and Applications , 2012, IEEE transactions on pattern analysis and machine intelligence.

[23]  Jiawei Han,et al.  Spectral Regression for Dimensionality Reduction , 2007 .

[24]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.

[25]  Wei Jia,et al.  Discriminant sparse neighborhood preserving embedding for face recognition , 2012, Pattern Recognit..

[26]  S. Mallat A wavelet tour of signal processing , 1998 .

[27]  Lei Wang,et al.  Global and Local Structure Preservation for Feature Selection , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[28]  D. Donoho For most large underdetermined systems of linear equations the minimal 𝓁1‐norm solution is also the sparsest solution , 2006 .

[29]  Zhong Jin,et al.  Face recognition using discriminant sparsity neighborhood preserving embedding , 2012, Knowl. Based Syst..

[30]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[31]  I. Jolliffe Principal Component Analysis , 2002 .

[32]  Pascal Frossard,et al.  Dictionary Learning , 2011, IEEE Signal Processing Magazine.

[33]  M. Elad,et al.  $rm K$-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation , 2006, IEEE Transactions on Signal Processing.

[34]  Stephen P. Boyd,et al.  An Interior-Point Method for Large-Scale $\ell_1$-Regularized Least Squares , 2007, IEEE Journal of Selected Topics in Signal Processing.

[35]  Jiawei Han,et al.  Semi-supervised Discriminant Analysis , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[36]  Jieping Ye,et al.  Integrating Global and Local Structures: A Least Squares Framework for Dimensionality Reduction , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[37]  Lei Zhang,et al.  Sparse representation or collaborative representation: Which helps face recognition? , 2011, 2011 International Conference on Computer Vision.

[38]  Bin Yu,et al.  Model Selection and the Principle of Minimum Description Length , 2001 .

[39]  Stphane Mallat,et al.  A Wavelet Tour of Signal Processing, Third Edition: The Sparse Way , 2008 .

[40]  Pingkun Yan,et al.  Alternatively Constrained Dictionary Learning For Image Superresolution , 2014, IEEE Transactions on Cybernetics.

[41]  Fatih Murat Porikli,et al.  Classification and Boosting with Multiple Collaborative Representations , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[42]  Yulong Wang,et al.  Sparse Coding From a Bayesian Perspective , 2013, IEEE Transactions on Neural Networks and Learning Systems.

[43]  Bernhard Schölkopf,et al.  Kernel Principal Component Analysis , 1997, ICANN.

[44]  Stephen J. Wright,et al.  Computational Methods for Sparse Solution of Linear Inverse Problems , 2010, Proceedings of the IEEE.

[45]  Michael Elad,et al.  Image Denoising Via Sparse and Redundant Representations Over Learned Dictionaries , 2006, IEEE Transactions on Image Processing.

[46]  Michael A. Saunders,et al.  Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..

[47]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[48]  B. Scholkopf,et al.  Fisher discriminant analysis with kernels , 1999, Neural Networks for Signal Processing IX: Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No.98TH8468).

[49]  Emmanuel J. Candès,et al.  Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.

[50]  Allen Y. Yang,et al.  Robust Face Recognition via Sparse Representation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[51]  Heeyoul Choi,et al.  Robust kernel Isomap , 2007, Pattern Recognit..

[52]  Xiaoyang Tan,et al.  Sparsity preserving discriminant analysis for single training image face recognition , 2010, Pattern Recognit. Lett..

[53]  Yuxiao Hu,et al.  Face recognition using Laplacianfaces , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[54]  R. Vidal,et al.  Sparse Subspace Clustering: Algorithm, Theory, and Applications. , 2013, IEEE transactions on pattern analysis and machine intelligence.

[55]  Anil K. Jain,et al.  Dimensionality reduction using genetic algorithms , 2000, IEEE Trans. Evol. Comput..

[56]  Xiaoyang Tan,et al.  Pattern Recognition , 2016, Communications in Computer and Information Science.

[57]  Jiawei Han,et al.  Spectral Regression for Efficient Regularized Subspace Learning , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[58]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .

[59]  Thomas S. Huang Locally Linear Embedded Eigenspace Analysis , 2005 .

[60]  Huan Liu,et al.  Efficient Feature Selection via Analysis of Relevance and Redundancy , 2004, J. Mach. Learn. Res..

[61]  Shuicheng Yan,et al.  Learning With $\ell ^{1}$-Graph for Image Analysis , 2010, IEEE Transactions on Image Processing.

[62]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[63]  Pingkun Yan,et al.  Image Super-Resolution Via Double Sparsity Regularized Manifold Learning , 2013, IEEE Transactions on Circuits and Systems for Video Technology.

[64]  Pascal Frossard,et al.  Dictionary learning: What is the right representation for my signal? , 2011 .