Orthogonal least squares regression for feature extraction

In many data mining applications, dimensionality reduction is a primary technique to map high-dimensional data to a lower dimensional space. In order to preserve more local structure information, we propose a novel orthogonal least squares regression model for feature extraction in this paper. The main contributions of this paper are shown as follows: first, the new least squares regression method is constructed under the orthogonal constraint which can preserve more discriminant information in the subspace. Second, the optimization problem of classical least squares regression can be solved easily. However the proposed objective function is an unbalanced orthogonal procrustes problem, it is so difficult to obtain the solution that we present a novel iterative optimization algorithm to obtain the optimal solution. The last one, we also provide a proof of the convergence for our iterative algorithm. Additionally, experimental results show that we obtain a global optimal solution through our iterative algorithm even though the optimization problem is a non-convex problem. Both theoretical analysis and empirical studies demonstrate that our method can more effectively reduce the data dimensionality than conventional methods.

[1]  Weiguo Fan,et al.  Effective and efficient dimensionality reduction for large-scale and streaming data preprocessing , 2006, IEEE Transactions on Knowledge and Data Engineering.

[2]  Ja-Chen Lin,et al.  A new LDA-based face recognition system which can solve the small sample size problem , 1998, Pattern Recognit..

[3]  Feiping Nie,et al.  Extracting the optimal dimensionality for local tensor discriminant analysis , 2009, Pattern Recognit..

[4]  Feiping Nie,et al.  Discriminative Least Squares Regression for Multiclass Classification and Feature Selection , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[5]  David J. Kriegman,et al.  Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection , 1996, ECCV.

[6]  J KriegmanDavid,et al.  Eigenfaces vs. Fisherfaces , 1997 .

[7]  Shiming Xiang,et al.  Retargeted Least Squares Regression Algorithm , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[8]  Chao Wang,et al.  Supervised feature extraction based on orthogonal discriminant projection , 2009, Neurocomputing.

[9]  J. Friedman Regularized Discriminant Analysis , 1989 .

[10]  Nittaya Kerdprasop,et al.  An Empirical Study of Dimensionality Reduction Methods for Biometric Recognition , 2014, 2014 7th International Conference on Security Technology.

[11]  Feiping Nie,et al.  Trace Ratio Problem Revisited , 2009, IEEE Transactions on Neural Networks.

[12]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[13]  Jingjing Liu,et al.  Enhanced fisher discriminant criterion for image recognition , 2012, Pattern Recognit..

[14]  Jing Peng,et al.  SVM vs regularized least squares classification , 2004, ICPR 2004.

[15]  Hua Yu,et al.  A direct LDA algorithm for high-dimensional data - with application to face recognition , 2001, Pattern Recognit..

[16]  Jiawei Han,et al.  Orthogonal Laplacianfaces for Face Recognition , 2006, IEEE Transactions on Image Processing.

[17]  Yuxiao Hu,et al.  Face recognition using Laplacianfaces , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Feiping Nie,et al.  Flexible Shift-Invariant Locality and Globality Preserving Projections , 2014, ECML/PKDD.

[19]  Hong Qiao,et al.  Scatter Balance: An Angle-Based Supervised Dimensionality Reduction , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[20]  T. Poggio,et al.  Regularized Least-Squares Classification 133 In practice , although , 2007 .

[21]  Changyin Sun,et al.  A regularized least square based discriminative projections for feature extraction , 2016, Neurocomputing.

[22]  Feiping Nie,et al.  Neighborhood MinMax Projections , 2007, IJCAI.

[23]  Yousef Saad,et al.  Orthogonal neighborhood preserving projections , 2005, Fifth IEEE International Conference on Data Mining (ICDM'05).

[24]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[25]  Dirk L. Knol,et al.  Orthogonal rotations to maximal agreement for two or more matrices of different column orders , 1984 .

[26]  Feiping Nie,et al.  Semi-supervised orthogonal discriminant analysis via label propagation , 2009, Pattern Recognit..

[27]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.

[28]  Christopher M. Bishop,et al.  Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .

[29]  Hailin Zhang,et al.  Discriminative sparsity preserving projections for image recognition , 2015, Pattern Recognit..

[30]  Feiping Nie,et al.  Orthogonal locality minimizing globality maximizing projections for feature extraction , 2009 .

[31]  Tao Jiang,et al.  Efficient and robust feature extraction by maximum margin criterion , 2003, IEEE Transactions on Neural Networks.

[32]  Jieping Ye,et al.  Least squares linear discriminant analysis , 2007, ICML '07.

[33]  Feiping Nie,et al.  Orthogonal vs. uncorrelated least squares discriminant analysis for feature extraction , 2012, Pattern Recognit. Lett..

[34]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .

[35]  Yousef Saad,et al.  Orthogonal Neighborhood Preserving Projections: A Projection-Based Dimensionality Reduction Technique , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[36]  Zhenyue Zhang,et al.  Successive projection method for solving the unbalanced Procrustes problem , 2006 .

[37]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[38]  Jieping Ye,et al.  Characterization of a Family of Algorithms for Generalized Discriminant Analysis on Undersampled Problems , 2005, J. Mach. Learn. Res..

[39]  R. Fisher THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMS , 1936 .