Neighborhood Selection for Dimensionality Reduction

Though a great deal of research work has been devoted to the development of dimensionality reduction algorithms, the problem is still open. The most recent and effective techniques, assuming datasets drawn from an underlying low dimensional manifold embedded into an high dimensional space, look for “small enough” neighborhoods which should represent the underlying manifold portion. Unfortunately, neighborhood selection is an open problem, for the presence of noise, outliers, points not uniformly distributed, and to unexpected high manifold curvatures, causing the inclusion of geodesically distant points in the same neighborhood. In this paper we describe our neighborhood selection algorithm, called ONeS; it exploits both distance and angular information to form neighborhoods containing nearby points that share a common local structure in terms of curvature. The reported experimental results show the enhanced quality of the neighborhoods computed by ONeS w.r.t. the commonly used k-neighborhoods solely employing the euclidean distance.

[1]  Alessandro Rozza,et al.  DANCo: An intrinsic dimensionality estimator exploiting angle and norm concentration , 2014, Pattern Recognit..

[2]  Hong Qiao,et al.  An improved local tangent space alignment method for manifold learning , 2011, Pattern Recognit. Lett..

[3]  Li Jing,et al.  Selection of the Suitable Parameter Value for ISOMAP , 2011, J. Softw..

[4]  Hongyuan Zha,et al.  Principal Manifolds and Nonlinear Dimension Reduction via Local Tangent Space Alignment , 2002, ArXiv.

[5]  Alfred O. Hero,et al.  Determining Intrinsic Dimension and Entropy of High-Dimensional Shape Spaces , 2006, Statistics and Analysis of Shapes.

[6]  Jia Wei,et al.  Adaptive neighborhood selection for manifold learning , 2008, 2008 International Conference on Machine Learning and Cybernetics.

[7]  Jianping Yin,et al.  Robust local tangent space alignment via iterative weighted PCA , 2011, Neurocomputing.

[8]  Jing Wang,et al.  Extended local tangent space alignment for classification , 2012, Neurocomputing.

[9]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[10]  Matthias Hein,et al.  Intrinsic Dimensionality Estimation of Submanifolds in Euclidean space , 2005, ICML 2005.

[11]  Peter J. Bickel,et al.  Maximum Likelihood Estimation of Intrinsic Dimension , 2004, NIPS.

[12]  H. Zha,et al.  Principal manifolds and nonlinear dimensionality reduction via tangent space alignment , 2004, SIAM J. Sci. Comput..

[13]  Alessandro Rozza,et al.  Novel high intrinsic dimensionality estimators , 2012, Machine Learning.

[14]  I. Jolliffe Principal Component Analysis , 2002 .

[15]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[16]  Tony R. Martinez,et al.  Tangent space guided intelligent neighbor finding , 2011, The 2011 International Joint Conference on Neural Networks.

[17]  Matthias Hein,et al.  Intrinsic dimensionality estimation of submanifolds in Rd , 2005, ICML.

[18]  Joshua B. Tenenbaum,et al.  Global Versus Local Methods in Nonlinear Dimensionality Reduction , 2002, NIPS.

[19]  Hongyuan Zha,et al.  Adaptive Manifold Learning , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  Jing Wang Improve local tangent space alignment using various dimensional local coordinates , 2008, Neurocomputing.

[21]  Tony R. Martinez,et al.  Robust manifold learning with CycleCut , 2012, Connect. Sci..

[22]  L. Rosasco,et al.  Multiscale geometric methods for data sets I: Multiscale SVD, noise and curvature , 2017 .

[23]  Jeanny Hérault,et al.  Curvilinear component analysis: a self-organizing neural network for nonlinear mapping of data sets , 1997, IEEE Trans. Neural Networks.