Spectral feature scaling method for supervised dimensionality reduction

Spectral dimensionality reduction methods enable linear separations of complex data with high-dimensional features in a reduced space. However, these methods do not always give the desired results due to irregularities or uncertainties of the data. Thus, we consider aggressively modifying the scales of the features to obtain the desired classification. Using prior knowledge on the labels of partial samples to specify the Fiedler vector, we formulate an eigenvalue problem of a linear matrix pencil whose eigenvector has the feature scaling factors. The resulting factors can modify the features of entire samples to form clusters in the reduced space, according to the known labels. In this study, we propose new dimensionality reduction methods supervised using the feature scaling associated with the spectral clustering. Numerical experiments show that the proposed methods outperform well-established supervised methods for toy problems with more samples than features, and are more robust regarding clustering than existing methods. Also, the proposed methods outperform existing methods regarding classification for real-world problems with more features than samples of gene expression profiles of cancer diseases. Furthermore, the feature scaling tends to improve the clustering and classification accuracies of existing unsupervised methods, as the proportion of training data increases.

[1]  Feiping Nie,et al.  A unified framework for semi-supervised dimensionality reduction , 2008, Pattern Recognit..

[2]  Yan Cui,et al.  A novel supervised dimensionality reduction algorithm: Graph-based Fisher analysis , 2012, Pattern Recognit..

[3]  A. Isaksson,et al.  Cross-validation and bootstrapping are unreliable in small sample classification , 2008, Pattern Recognit. Lett..

[4]  Rong Jin,et al.  Distance Metric Learning: A Comprehensive Survey , 2006 .

[5]  David J. Kriegman,et al.  Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection , 1996, ECCV.

[6]  M. Fiedler Algebraic connectivity of graphs , 1973 .

[7]  C. Fombrun,et al.  Social Network Analysis For Organizations , 1979 .

[8]  William M. Rand,et al.  Objective Criteria for the Evaluation of Clustering Methods , 1971 .

[9]  Jitendra Malik,et al.  Normalized cuts and image segmentation , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[10]  S. Drăghici,et al.  Analysis of microarray experiments of gene expression profiling. , 2006, American journal of obstetrics and gynecology.

[11]  Dan Klein,et al.  Spectral Learning , 2003, IJCAI.

[12]  Inderjit S. Dhillon,et al.  Clustering with Bregman Divergences , 2005, J. Mach. Learn. Res..

[13]  Xiaofei He,et al.  Locality Preserving Projections , 2003, NIPS.

[14]  Inderjit S. Dhillon,et al.  Semi-supervised graph clustering: a kernel approach , 2005, ICML '05.

[15]  Peter C. Jurs,et al.  Computational Methods for the Analysis of Chemical Sensor Array Data from Volatile Analytes , 2000 .

[16]  Fan Chung,et al.  Spectral Graph Theory , 1996 .

[17]  Shang-Hua Teng,et al.  Nearly-Linear Time Algorithms for Preconditioning and Solving Symmetric, Diagonally Dominant Linear Systems , 2006, SIAM J. Matrix Anal. Appl..

[18]  Kazuo Murota,et al.  An Algorithm for the Generalized Eigenvalue Problem for Nonsquare Matrix Pencils by Minimal Perturbation Approach , 2016, SIAM J. Matrix Anal. Appl..

[19]  Masashi Sugiyama,et al.  Dimensionality Reduction of Multimodal Labeled Data by Local Fisher Discriminant Analysis , 2007, J. Mach. Learn. Res..

[20]  Peter Tiño,et al.  Indefinite Proximity Learning: A Review , 2015, Neural Computation.

[21]  E. M. Sá Czechoslovak Mathematical Journal , 2016 .

[22]  Amos Storkey,et al.  Advances in Neural Information Processing Systems 20 , 2007 .

[23]  Eran Stark,et al.  Spike sorting: Bayesian clustering of non-stationary data , 2004, Journal of Neuroscience Methods.

[24]  Tarek S. Sobh Wired and wireless intrusion detection system: Classifications, good characteristics and state-of-the-art , 2006, Comput. Stand. Interfaces.

[25]  Joydeep Ghosh,et al.  Cluster Ensembles --- A Knowledge Reuse Framework for Combining Multiple Partitions , 2002, J. Mach. Learn. Res..

[26]  Michael I. Jordan,et al.  Learning Spectral Clustering , 2003, NIPS.

[27]  David J. Kriegman,et al.  Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection , 1996, ECCV.

[28]  S. A. Sherman,et al.  Providence , 1906 .

[29]  Xuelong Li,et al.  Locality Adaptive Discriminant Analysis , 2017, IJCAI.