On Nonlinear Dimensionality Reduction, Linear Smoothing and Autoencoding

We develop theory for nonlinear dimensionality reduction (NLDR). A number of NLDR methods have been developed, but there is limited understanding of how these methods work and the relationships between them. There is limited basis for using existing NLDR theory for deriving new algorithms. We provide a novel framework for analysis of NLDR via a connection to the statistical theory of linear smoothers. This allows us to both understand existing methods and derive new ones. We use this connection to smoothing to show that asymptotically, existing NLDR methods correspond to discrete approximations of the solutions of sets of differential equations given a boundary condition. In particular, we can characterize many existing methods in terms of just three limiting differential operators and boundary conditions. Our theory also provides a way to assert that one method is preferable to another; indeed, we show Local Tangent Space Alignment is superior within a class of methods that assume a global coordinate chart defines an isometric embedding of the manifold.

[1]  R. Tibshirani,et al.  Linear Smoothers and Additive Models , 1989 .

[2]  M. Wand,et al.  Multivariate Locally Weighted Least Squares Regression , 1994 .

[3]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[4]  D. Donoho,et al.  Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data , 2003, Proceedings of the National Academy of Sciences of the United States of America.

[5]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.

[6]  张振跃,et al.  Principal Manifolds and Nonlinear Dimensionality Reduction via Tangent Space Alignment , 2004 .

[7]  R. Coifman,et al.  Diffusion Wavelets , 2004 .

[8]  Kilian Q. Weinberger,et al.  Unsupervised Learning of Image Manifolds by Semidefinite Programming , 2004, CVPR.

[9]  Jing Wang,et al.  MLLE: Modified Locally Linear Embedding Using Multiple Weights , 2006, NIPS.

[10]  Mikhail Belkin,et al.  Convergence of Laplacian Eigenmaps , 2006, NIPS.

[11]  Stéphane Lafon,et al.  Diffusion maps , 2006 .

[12]  Ulrike von Luxburg,et al.  Graph Laplacians and their Convergence on Random Neighborhood Graphs , 2006, J. Mach. Learn. Res..

[13]  B. Schölkopf,et al.  MLLE: Modified Locally Linear Embedding Using Multiple Weights , 2007 .

[14]  Yaacov Ritov,et al.  LDR-LLE: LLE with Low-Dimensional Neighborhood Representation , 2008, ISVC.

[15]  Mikhail Belkin,et al.  Consistency of spectral clustering , 2008, 0804.0678.

[16]  Ling Huang,et al.  An Analysis of the Convergence of Graph Laplacians , 2010, ICML.

[17]  Hau-Tieng Wu,et al.  Local Linear Regression on Manifolds and Its Geometric Interpretation , 2012, 1201.0327.

[18]  Amit Singer,et al.  Spectral Convergence of the connection Laplacian from random samples , 2013, 1306.1587.

[19]  Zhengming Ma,et al.  On the Equivalence of HLLE and LTSA , 2018, IEEE Transactions on Cybernetics.

[20]  Tyrus Berry,et al.  Consistent manifold representation for topological data analysis , 2016, Foundations of Data Science.