Nonlinear Dimensionality Reduction by Locally Linear Inlaying

High-dimensional data is involved in many fields of information processing. However, sometimes, the intrinsic structures of these data can be described by a few degrees of freedom. To discover these degrees of freedom or the low-dimensional nonlinear manifold underlying a high-dimensional space, many manifold learning algorithms have been proposed. Here we describe a novel algorithm, locally linear inlaying (LLI), which combines simple geometric intuitions and rigorously established optimality to compute the global embedding of a nonlinear manifold. Using a divide-and-conquer strategy, LLI gains some advantages in itself. First, its time complexity is linear in the number of data points, and hence LLI can be implemented efficiently. Second, LLI overcomes problems caused by the nonuniform sample distribution. Third, unlike existing algorithms such as isometric feature mapping (Isomap), local tangent space alignment (LTSA), and locally linear coordination (LLC), LLI is robust to noise. In addition, to evaluate the embedding results quantitatively, two criteria based on information theory and Kolmogorov complexity theory, respectively, are proposed. Furthermore, we demonstrated the efficiency and effectiveness of our proposal by synthetic and real-world data sets.

[1]  Lucian Ilie,et al.  The Lempel--Ziv Complexity of Fixed Points of Morphisms , 2006, SIAM J. Discret. Math..

[2]  Teuvo Kohonen,et al.  Self-Organizing Maps , 2010 .

[3]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[4]  Abhi Shelat,et al.  Approximating the smallest grammar: Kolmogorov complexity in natural models , 2002, STOC '02.

[5]  Hongbin Zha,et al.  Riemannian Manifold Learning for Nonlinear Dimensionality Reduction , 2006, ECCV.

[6]  Jeanny Hérault,et al.  Curvilinear component analysis: a self-organizing neural network for nonlinear mapping of data sets , 1997, IEEE Trans. Neural Networks.

[7]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[8]  T. Hastie,et al.  Principal Curves , 2007 .

[9]  George W. Irwin,et al.  RBF principal manifolds for process monitoring , 1999, IEEE Trans. Neural Networks.

[10]  Ross T. Whitaker,et al.  Robust non-linear dimensionality reduction using successive 1-dimensional Laplacian Eigenmaps , 2007, ICML '07.

[11]  C. Tsallis Possible generalization of Boltzmann-Gibbs statistics , 1988 .

[12]  Teofilo F. GONZALEZ,et al.  Clustering to Minimize the Maximum Intercluster Distance , 1985, Theor. Comput. Sci..

[13]  Haixian Wang,et al.  Locality-Preserved Maximum Information Projection , 2008, IEEE Transactions on Neural Networks.

[14]  Joshua B. Tenenbaum,et al.  Global Versus Local Methods in Nonlinear Dimensionality Reduction , 2002, NIPS.

[15]  Masakazu Kojima,et al.  Exact Solutions of Some Nonconvex Quadratic Optimization Problems via SDP and SOCP Relaxations , 2003, Comput. Optim. Appl..

[16]  Dit-Yan Yeung,et al.  Robust locally linear embedding , 2006, Pattern Recognit..

[17]  David S. Johnson,et al.  Computers and Intractability: A Guide to the Theory of NP-Completeness , 1978 .

[18]  Masahiro Tanaka Mixture of Probabilistic Factor Analysis Model and Its Applications , 2001, ICANN.

[19]  Thomas Martinetz,et al.  Topology representing networks , 1994, Neural Networks.

[20]  Hou Yuexian,et al.  Improving classification precision by implicit kernels motivated by manifold learning , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[21]  David S. Watkins,et al.  Fundamentals of Matrix Computations: Watkins/Fundamentals of Matrix Computations , 2005 .

[22]  Diego Vidaurre,et al.  A Quick Assessment of Topology Preservation for SOM Structures , 2007, IEEE Transactions on Neural Networks.

[23]  H. Zha,et al.  Principal manifolds and nonlinear dimensionality reduction via tangent space alignment , 2004, SIAM J. Sci. Comput..

[24]  Jan Kybic,et al.  High-Dimensional Entropy Estimation for Finite Accuracy Data: R-NN Entropy Estimator , 2007, IPMI.

[25]  Yee Whye Teh,et al.  Automatic Alignment of Local Representations , 2002, NIPS.

[26]  Heng Tao Shen,et al.  Principal Component Analysis , 2009, Encyclopedia of Biometrics.

[27]  B. Ripley,et al.  Robust Statistics , 2018, Encyclopedia of Mathematical Geosciences.

[28]  Ming Li,et al.  An Introduction to Kolmogorov Complexity and Its Applications , 2019, Texts in Computer Science.

[29]  David S. Watkins,et al.  Fundamentals of matrix computations , 1991 .

[30]  Jin Hyeong Park,et al.  Local smoothing for manifold learning , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[31]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[32]  B. Schölkopf,et al.  MLLE: Modified Locally Linear Embedding Using Multiple Weights , 2007 .

[33]  Zhang Yi,et al.  Dynamics of Generalized PCA and MCA Learning Algorithms , 2007, IEEE Transactions on Neural Networks.

[34]  Mikhail Belkin,et al.  Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering , 2001, NIPS.

[35]  Peter J. Huber,et al.  Robust Statistics , 2005, Wiley Series in Probability and Statistics.

[36]  Geoffrey E. Hinton,et al.  Global Coordination of Local Linear Models , 2001, NIPS.