Geodesic entropic graphs for dimension and entropy estimation in manifold learning

In the manifold learning problem, one seeks to discover a smooth low dimensional surface, i.e., a manifold embedded in a higher dimensional linear vector space, based on a set of measured sample points on the surface. In this paper, we consider the closely related problem of estimating the manifold's intrinsic dimension and the intrinsic entropy of the sample points. Specifically, we view the sample points as realizations of an unknown multivariate density supported on an unknown smooth manifold. We introduce a novel geometric approach based on entropic graph methods. Although the theory presented applies to this general class of graphs, we focus on the geodesic-minimal-spanning-tree (GMST) to obtaining asymptotically consistent estimates of the manifold dimension and the Re/spl acute/nyi /spl alpha/-entropy of the sample density on the manifold. The GMST approach is striking in its simplicity and does not require reconstruction of the manifold or estimation of the multivariate density of the samples. The GMST method simply constructs a minimal spanning tree (MST) sequence using a geodesic edge matrix and uses the overall lengths of the MSTs to simultaneously estimate manifold dimension and entropy. We illustrate the GMST approach on standard synthetic manifolds as well as on real data sets consisting of images of faces.

[1]  David J. Kriegman,et al.  From Few to Many: Illumination Cone Models for Face Recognition under Variable Lighting and Pose , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Anil K. Jain,et al.  Algorithms for Clustering Data , 1988 .

[3]  I. Hassan Embedded , 2005, The Cyber Security Handbook.

[4]  Aaas News,et al.  Book Reviews , 1893, Buffalo Medical and Surgical Journal.

[5]  Robert P. W. Duin,et al.  An Evaluation of Intrinsic Dimensionality Estimators , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[6]  David L. Neuhoff,et al.  On the asymptotic distribution of the errors in vector quantization , 1996, IEEE Trans. Inf. Theory.

[7]  Vin de Silva,et al.  Unsupervised Learning of Curved Manifolds , 2003 .

[8]  T. Willmore AN INTRODUCTION TO DIFFERENTIABLE MANIFOLDS AND RIEMANNIAN GEOMETRY (Second edition) , 1987 .

[9]  I. Holopainen Riemannian Geometry , 1927, Nature.

[10]  W. Boothby An introduction to differentiable manifolds and Riemannian geometry , 1975 .

[11]  Yachen Lin,et al.  Geometric Data Analysis: An Empirical Approach to Dimensionality Reduction and the Study of Patterns , 2002, Technometrics.

[12]  S. Osher,et al.  Solving variational problems and partial differential equations mapping into general target manifolds , 2004 .

[13]  Kenneth S. Alexander,et al.  The RSW theorem for continuum percolation and the CLT for Euclidean minimal spanning trees , 1996 .

[14]  X. Huo,et al.  LOCAL LINEAR PROJECTION (LLP) , 2002 .

[15]  Alfred O. Hero,et al.  Convergence rates of minimal graphs with random vertices , 2002 .

[16]  I. Csiszár Generalized Cutoff Rates and Renyi's Information Measures , 1993, Proceedings. IEEE International Symposium on Information Theory.

[17]  Allen Gersho,et al.  Asymptotically optimal block quantization , 1979, IEEE Trans. Inf. Theory.

[18]  H. Piaggio Differential Geometry of Curves and Surfaces , 1952, Nature.

[19]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[20]  A. Hero,et al.  Estimation of Renyi information divergence via pruned minimal spanning trees , 1999, Proceedings of the IEEE Signal Processing Workshop on Higher-Order Statistics. SPW-HOS '99.

[21]  Balázs Kégl,et al.  Intrinsic Dimension Estimation Using Packing Numbers , 2002, NIPS.

[22]  D. Bertsimas,et al.  The Minimum Spanning Tree Constant in Geometrical Probability and Under the Independent Model: A Unified Approach , 1992 .

[23]  J. Beardwood,et al.  The shortest path through many points , 1959, Mathematical Proceedings of the Cambridge Philosophical Society.

[24]  Anil K. Jain,et al.  An Intrinsic Dimensionality Estimator from Near-Neighbor Information , 1979, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[26]  Mikhail Belkin,et al.  Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering , 2001, NIPS.

[27]  Imre Csiszár Generalized cutoff rates and Renyi's information measures , 1995, IEEE Trans. Inf. Theory.

[28]  Alfred O. Hero,et al.  Asymptotic theory of greedy approximations to minimal k-point random graphs , 1999, IEEE Trans. Inf. Theory.

[29]  D. Bertsimas,et al.  An asymptotic determination of the minimum spanning tree and minimum matching constants in geometrical probability , 1990 .

[30]  F. Mémoli,et al.  Fast computation of weighted distance functions and geodesics on implicit hyper-surfaces: 730 , 2001 .

[31]  J. Yukich Probability theory of classical Euclidean optimization problems , 1998 .

[32]  D. Donoho,et al.  Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data , 2003, Proceedings of the National Academy of Sciences of the United States of America.

[33]  Alfred O. Hero,et al.  Applications of entropic spanning graphs , 2002, IEEE Signal Process. Mag..

[34]  Joshua B. Tenenbaum,et al.  Global Versus Local Methods in Nonlinear Dimensionality Reduction , 2002, NIPS.

[35]  A. Hero,et al.  Entropic graphs for manifold learning , 2003, The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003.

[36]  V. Koltchinskii Empirical geometry of multivariate data: a deconvolution approach , 2000 .

[37]  J. Steele Probability theory and combinatorial optimization , 1987 .

[38]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.