The Impact of Supervised Manifold Learning on Structure Preserving and Classification Error: A Theoretical Study

In recent years, a variety of supervised manifold learning techniques have been proposed to outperform their unsupervised alternative versions in terms of classification accuracy and data structure capturing. Some dissimilarity measures have been used in these techniques to guide the dimensionality reduction process. Their good performance was empirically demonstrated; however, the relevant analysis is still missing. This paper contributes to a theoretical analysis on a) how dissimilarity measures affect maintaining manifold neighbourhood structure and b) how supervised manifold learning techniques could contribute to the reduction of classification error. This paper also provides a cross-comparison between supervised and unsupervised manifold learning approaches in terms of structure capturing using Kendall’s Tau coefficients and co-ranking matrices. Four different metrics (including three dissimilarity measures and Euclidean distance) have been considered along with manifold learning methods such as Isomap, ${t}$ -Stochastic Neighbour Embedding ( ${t}$ -SNE), and Laplacian Eigenmaps (LE), in two datasets: Breast Cancer and Swiss-Roll. This paper concludes that although the dissimilarity measures used in the manifold learning techniques can reduce classification error, they do not learn well or preserve the structure of the hidden manifold in the high dimensional space, but instead, they destroy the structure of the data. Based on the findings of this paper, it is advisable to use supervised manifold learning techniques as a pre-processing step in classification. In addition, it is not advisable to apply supervised manifold learning for visualization purposes since the two-dimensional representation using supervised manifold learning does not improve the preservation of data structure.

[1]  Hongyuan Zha,et al.  Principal manifolds and nonlinear dimensionality reduction via tangent space alignment , 2004, SIAM J. Sci. Comput..

[2]  Gangyao Kuang,et al.  Deep supervised t-SNE for SAR target recognition , 2017, 2017 2nd International Conference on Frontiers of Sensors Technologies (ICFST).

[3]  Sidan Du,et al.  Robust Hessian Locally Linear Embedding Techniques for High-Dimensional Data , 2016, Algorithms.

[4]  Michel Verleysen,et al.  Quality assessment of dimensionality reduction: Rank-based criteria , 2009, Neurocomputing.

[5]  Junghui Chen,et al.  Developments of two supervised maximum variance unfolding algorithms for process classification , 2016 .

[6]  Lori M. Bruce,et al.  Why principal component analysis is not an appropriate feature extraction method for hyperspectral data , 2003, IGARSS 2003. 2003 IEEE International Geoscience and Remote Sensing Symposium. Proceedings (IEEE Cat. No.03CH37477).

[7]  Xin Geng,et al.  Supervised nonlinear dimensionality reduction for visualization and classification , 2005, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[8]  Yi Peng,et al.  Nonlinear manifold learning for early warnings in financial markets , 2017, Eur. J. Oper. Res..

[9]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[10]  Maria-Florina Balcan,et al.  On a theory of learning with similarity functions , 2006, ICML.

[11]  Ann B. Lee,et al.  Diffusion maps and coarse-graining: a unified framework for dimensionality reduction, graph partitioning, and data set parameterization , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Hongsheng Li,et al.  Silhouette Analysis for Human Action Recognition Based on Supervised Temporal t-SNE and Incremental Learning , 2015, IEEE Transactions on Image Processing.

[13]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[14]  Dimitrios Gunopulos,et al.  Non-linear dimensionality reduction techniques for classification and visualization , 2002, KDD.

[15]  Kilian Q. Weinberger,et al.  Unsupervised Learning of Image Manifolds by Semidefinite Programming , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[16]  J. Kruskal Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis , 1964 .

[17]  Mikhail Belkin,et al.  Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering , 2001, NIPS.

[18]  Shi-qing Zhang,et al.  Enhanced supervised locally linear embedding , 2009, Pattern Recognit. Lett..

[19]  Kewei Chen,et al.  Visualizing Alzheimer's disease progression in low dimensional manifolds , 2019, Heliyon.

[20]  D. D. Ridder,et al.  Locally linear embedding for classification , 2002 .

[21]  Michael Kirby,et al.  Supervised Dimensionality Reduction and Visualization using Centroid-encoder , 2020, J. Mach. Learn. Res..

[22]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[23]  Alexander M. Bronstein,et al.  Nonlinear Dimensionality Reduction by Topologically Constrained Isometric Embedding , 2010, International Journal of Computer Vision.

[24]  Laureta Hajderanj,et al.  A New Supervised t-SNE with Dissimilarity Measure for Effective Data Visualization and Classification , 2019, ICSIE.

[25]  Zhengming Ma,et al.  Local Coordinates Alignment With Global Preservation for Dimensionality Reduction , 2013, IEEE Transactions on Neural Networks and Learning Systems.

[26]  Alan Julian Izenman,et al.  Modern Multivariate Statistical Techniques , 2008 .

[27]  张振跃,et al.  Principal Manifolds and Nonlinear Dimensionality Reduction via Tangent Space Alignment , 2004 .

[28]  H. Hotelling Analysis of a complex of statistical variables into principal components. , 1933 .

[29]  Samuel Kadoury Manifold Learning in Medical Imaging , 2018 .

[30]  Leland McInnes,et al.  UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction , 2018, ArXiv.

[31]  Ronald R. Coifman,et al.  Data Fusion and Multicue Data Matching by Diffusion Maps , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  Enrico Grisan,et al.  Single- and Multi-Distribution Dimensionality Reduction Approaches for a Better Data Structure Capturing , 2020, IEEE Access.