Graph embedding based feature selection

Usually many real datasets in pattern recognition applications contain a large quantity of noisy and redundant features that are irrelevant to the intrinsic characteristics of the dataset. The irrelevant features may seriously deteriorate the learning performance. Hence feature selection which aims to select the most informative features and to eliminate the irrelevant feature from the original dataset plays an important role in data mining, image recognition and microarray data analysis. In this paper, we developed a new feature selection technique based on the recently developed graph embedding framework for manifold learning. We propose a recursive feature elimination (RFE) method using feature score for identifying the optimal feature subset. One advantage of the RFE method is that it can successfully identify the nonlinear features based on manifold learning. The experimental results both on face dataset and microarray dataset verify the effectiveness and efficiency of the proposed method.

[1]  Zenglin Xu,et al.  Discriminative Semi-Supervised Feature Selection Via Manifold Regularization , 2009, IEEE Transactions on Neural Networks.

[2]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[3]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[4]  Michael T. Manry,et al.  Feature Selection Using a Piecewise Linear Network , 2006, IEEE Transactions on Neural Networks.

[5]  Jidong Zhao,et al.  Locality sensitive semi-supervised feature selection , 2008, Neurocomputing.

[6]  Deng Cai,et al.  Laplacian Score for Feature Selection , 2005, NIPS.

[7]  Jiawei Han,et al.  Orthogonal Laplacianfaces for Face Recognition , 2006, IEEE Transactions on Image Processing.

[8]  Elmar Wolfgang Lang,et al.  Unsupervised feature extraction via kernel subspace techniques , 2011, Neurocomputing.

[9]  Huiqing Liu,et al.  A comparative study on feature selection and classification methods using gene expression profiles and proteomic patterns. , 2002, Genome informatics. International Conference on Genome Informatics.

[10]  Wen Wen,et al.  Kernel based gene expression pattern discovery and its application on cancer classification , 2010, Neurocomputing.

[11]  Stephen Lin,et al.  Marginal Fisher Analysis and Its Variants for Human Gait Recognition and Content- Based Image Retrieval , 2007, IEEE Transactions on Image Processing.

[12]  Feiping Nie,et al.  Trace Ratio Criterion for Feature Selection , 2008, AAAI.

[13]  Feiping Nie,et al.  Efficient and Robust Feature Selection via Joint ℓ2, 1-Norms Minimization , 2010, NIPS.

[14]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[15]  Feiping Nie,et al.  Neighborhood MinMax Projections , 2007, IJCAI.

[16]  Avinash C. Kak,et al.  PCA versus LDA , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[17]  Min Han,et al.  Sparse kernel density estimations and its application in variable selection based on quadratic Renyi entropy , 2011, Neurocomputing.

[18]  Jiawei Han,et al.  Learning a Maximum Margin Subspace for Image Retrieval , 2008, IEEE Transactions on Knowledge and Data Engineering.

[19]  Enrique Alba,et al.  Gene selection in cancer classification using PSO/SVM and GA/SVM hybrid algorithms , 2007, 2007 IEEE Congress on Evolutionary Computation.

[20]  J. Mesirov,et al.  Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. , 1999, Science.

[21]  Jason Weston,et al.  Gene Selection for Cancer Classification using Support Vector Machines , 2002, Machine Learning.

[22]  Qinghua Hu,et al.  Feature evaluation and selection based on neighborhood soft margin , 2010, Neurocomputing.

[23]  Shuicheng Yan,et al.  Graph embedding: a general framework for dimensionality reduction , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[24]  Hiok Chai Quek,et al.  MCES: A Novel Monte Carlo Evaluative Selection Approach for Objective Feature Selections , 2007, IEEE Transactions on Neural Networks.