A graph-based approach for feature selection from higher order correlations

Graph technology emerges as an important topic in the field of data mining and machine learning community. The analysis of high-dimensional data is crucial to identify a smaller subset of features which are informative for classification and clustering. In this paper, an efficient graph feature selection method is proposed to render the analysis of high-dimensional data tractable. Here, the feature scores are calculated for obtaining the weights of the edges in the weighted graph to identify the optimal feature subset. One advantage of this method is that it can successfully identify the optimal features for machine learning. The experimental results on our dataset verify the effectiveness and efficiency of the proposed method.

[1]  Fan Chung,et al.  Spectral Graph Theory , 1996 .

[2]  Yiming Yang,et al.  A Comparative Study on Feature Selection in Text Categorization , 1997, ICML.

[3]  Jason Weston,et al.  Gene Selection for Cancer Classification using Support Vector Machines , 2002, Machine Learning.

[4]  Mikhail Belkin,et al.  Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering , 2001, NIPS.

[5]  Avinash C. Kak,et al.  PCA versus LDA , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[6]  Ron Kohavi,et al.  Wrappers for Feature Subset Selection , 1997, Artif. Intell..

[7]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[8]  Lei Wang,et al.  Feature Selection With Redundancy-Constrained Class Separability , 2010, IEEE Transactions on Neural Networks.

[9]  Hua Yu,et al.  A direct LDA algorithm for high-dimensional data - with application to face recognition , 2001, Pattern Recognit..

[10]  Yiming Yang,et al.  From Lasso regression to Feature vector machine , 2005, NIPS.

[11]  Alexander J. Smola,et al.  Learning with kernels , 1998 .

[12]  Vojislav Kecman,et al.  Semi-supervised learning from unbalanced labeled data: An improvement , 2006 .

[13]  Shuicheng Yan,et al.  Graph embedding: a general framework for dimensionality reduction , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[14]  Kohei Inoue,et al.  Dimensionality Reduction for Semi-supervised Face Recognition , 2005, FSKD.

[15]  David S. Johnson,et al.  Computers and Intractability: A Guide to the Theory of NP-Completeness , 1978 .

[16]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[17]  Ian T. Jolliffe,et al.  Principal Component Analysis , 1986, Springer Series in Statistics.

[18]  Hiok Chai Quek,et al.  MCES: A Novel Monte Carlo Evaluative Selection Approach for Objective Feature Selections , 2007, IEEE Transactions on Neural Networks.

[19]  Zoubin Ghahramani,et al.  Combining active learning and semi-supervised learning using Gaussian fields and harmonic functions , 2003, ICML 2003.

[20]  Bernhard Schölkopf,et al.  Learning with Local and Global Consistency , 2003, NIPS.

[21]  Michael T. Manry,et al.  Feature Selection Using a Piecewise Linear Network , 2006, IEEE Transactions on Neural Networks.

[22]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[23]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.