Principal Polynomial Analysis

This paper presents a new framework for manifold learning based on a sequence of principal polynomials that capture the possibly nonlinear nature of the data. The proposed Principal Polynomial Analysis (PPA) generalizes PCA by modeling the directions of maximal variance by means of curves, instead of straight lines. Contrarily to previous approaches, PPA reduces to performing simple univariate regressions, which makes it computationally feasible and robust. Moreover, PPA shows a number of interesting analytical properties. First, PPA is a volume-preserving map, which in turn guarantees the existence of the inverse. Second, such an inverse can be obtained in closed form. Invertibility is an important advantage over other learning methods, because it permits to understand the identified features in the input domain where the data has physical meaning. Moreover, it allows to evaluate the performance of dimensionality reduction in sensible (input-domain) units. Volume preservation also allows an easy computation of information theoretic quantities, such as the reduction in multi-information after the transform. Third, the analytical nature of PPA leads to a clear geometrical interpretation of the manifold: it allows the computation of Frenet-Serret frames (local features) and of generalized curvatures at any point of the space. And fourth, the analytical Jacobian allows the computation of the metric induced by the data, thus generalizing the Mahalanobis distance. These properties are demonstrated theoretically and illustrated experimentally. The performance of PPA is evaluated in dimensionality and redundancy reduction, in both synthetic and real datasets from the UCI repository.

[1]  T. Hastie Principal Curves and Surfaces , 1984 .

[2]  Yee Whye Teh,et al.  Automatic Alignment of Local Representations , 2002, NIPS.

[3]  Valero Laparra,et al.  Nonlinearities and Adaptation of Color Vision from Sequential Principal Curves Analysis , 2016, Neural Computation.

[4]  Paul Honeine,et al.  Preimage Problem in Kernel-Based Machine Learning , 2011, IEEE Signal Processing Magazine.

[5]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[6]  Jarkko Venna,et al.  Information Retrieval Perspective to Nonlinear Dimensionality Reduction for Data Visualization , 2010, J. Mach. Learn. Res..

[7]  Kaare Brandt Petersen,et al.  Kernel Multivariate Analysis Framework for Supervised Subspace Learning: A Tutorial on Linear and Kernel Multivariate Methods , 2013, IEEE Signal Processing Magazine.

[8]  Jesús Malo,et al.  The Role of Spatial Information in Disentangling the Irradiance–Reflectance–Transmittance Ambiguity , 2014, IEEE Transactions on Geoscience and Remote Sensing.

[9]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[10]  Matthias Scholz,et al.  Validation of Nonlinear PCA , 2012, Neural Processing Letters.

[11]  Michel Verleysen,et al.  Nonlinear Dimensionality Reduction , 2021, Computer Vision.

[12]  Joachim M. Buhmann,et al.  On Relevant Dimensions in Kernel Feature Spaces , 2008, J. Mach. Learn. Res..

[13]  Ahmed Bouridane,et al.  2D and 3D palmprint information, PCA and HMM for an improved person recognition performance , 2013, Integr. Comput. Aided Eng..

[14]  Jesús Malo,et al.  Linear transform for simultaneous diagonalization of covariance and perceptual metric matrix in image coding , 2003, Pattern Recognit..

[15]  Valero Laparra,et al.  Nonlinear data description with Principal Polynomial Analysis , 2012, 2012 IEEE International Workshop on Machine Learning for Signal Processing.

[16]  Valero Laparra,et al.  Principal polynomial analysis for remote sensing data processing , 2011, 2011 IEEE International Geoscience and Remote Sensing Symposium.

[17]  Gerhard Tutz,et al.  Local principal curves , 2005, Stat. Comput..

[18]  Joachim Selbig,et al.  Non-linear PCA: a missing data approach , 2005, Bioinform..

[19]  Valero Laparra,et al.  Iterative Gaussianization: From ICA to Random Rotations , 2011, IEEE Transactions on Neural Networks.

[20]  P. Delicado Another Look at Principal Curves and Surfaces , 2001 .

[21]  John Suckling,et al.  Intrinsic Curvature: a Marker of Millimeter-Scale Tangential cortico-Cortical Connectivity? , 2011, Int. J. Neural Syst..

[22]  Jochen Einbeck,et al.  Data Compression and Regression through Local Principal Curves and Surfaces , 2010, Int. J. Neural Syst..

[23]  Eero P. Simoncelli,et al.  Nonlinear Extraction of Independent Components of Natural Images Using Radial Gaussianization , 2009, Neural Computation.

[24]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[25]  Kilian Q. Weinberger,et al.  Unsupervised Learning of Image Manifolds by Semidefinite Programming , 2004, CVPR.

[26]  Ezequiel López-Rubio,et al.  Foreground Detection in Video Sequences with Probabilistic Self-Organizing Maps , 2011, Int. J. Neural Syst..

[27]  Chandan Chakraborty,et al.  Application of Higher Order cumulant Features for Cardiac Health Diagnosis using ECG signals , 2013, Int. J. Neural Syst..

[28]  Matthew Brand,et al.  Charting a Manifold , 2002, NIPS.

[29]  Matthias Scholz,et al.  Nonlinear Principal Component Analysis: Neural Network Models and Applications , 2008 .

[30]  J. David Logan,et al.  An Introduction to Nonlinear Partial Differential Equations , 1994 .

[31]  David G. Stork,et al.  Pattern Classification (2nd ed.) , 1999 .

[32]  P. Mahalanobis On the generalized distance in statistics , 1936 .

[33]  Geoffrey E. Hinton,et al.  Global Coordination of Local Linear Models , 2001, NIPS.

[34]  Valero Laparra,et al.  Divisive normalization image quality metric revisited. , 2010, Journal of the Optical Society of America. A, Optics, image science, and vision.

[35]  Deniz Erdogmus,et al.  Locally Defined Principal Curves and Surfaces , 2011, J. Mach. Learn. Res..

[36]  Inés María Galván,et al.  Recursive Discriminant Regression Analysis to Find Homogeneous Groups , 2011, Int. J. Neural Syst..

[37]  Ben J. A. Kröse,et al.  Coordinating Principal Component Analyzers , 2002, ICANN.

[38]  Ian T. Jolliffe,et al.  Principal Component Analysis , 2002, International Encyclopedia of Statistical Science.

[39]  Robin Sibson,et al.  What is projection pursuit , 1987 .

[40]  B. Dubrovin,et al.  Modern geometry--methods and applications , 1984 .

[41]  Ezequiel López-Rubio,et al.  Bregman Divergences for Growing Hierarchical Self-Organizing Networks , 2014, Int. J. Neural Syst..

[42]  D. J. Donnell,et al.  Analysis of Additive Dependencies and Concurvities Using Smallest Additive Principal Components , 1994 .

[43]  Hojjat Adeli,et al.  Principal Component Analysis-Enhanced Cosine Radial Basis Function Neural Network for Robust Epilepsy and Seizure Detection , 2008, IEEE Transactions on Biomedical Engineering.

[44]  Christopher J. C. Burges,et al.  Geometry and invariance in kernel based methods , 1999 .

[45]  Eero P. Simoncelli,et al.  Nonlinear image representation for efficient perceptual coding , 2006, IEEE Transactions on Image Processing.

[46]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[47]  M. Kramer Nonlinear principal component analysis using autoassociative neural networks , 1991 .

[48]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[49]  Ulrik Söderström,et al.  Reconstruction of occluded facial images using asymmetrical Principal Component Analysis , 2011, 2011 18th International Conference on Systems, Signals and Image Processing.

[50]  Dewang Chen,et al.  A Riemannian Distance Approach for Constructing Principal Curves , 2010, Int. J. Neural Syst..

[51]  T. Hastie,et al.  Principal Curves , 2007 .

[52]  Adam Krzyzak,et al.  Piecewise Linear Skeletonization Using Principal Curves , 2002, IEEE Trans. Pattern Anal. Mach. Intell..