Statistical mechanics of learning orthogonal signals for general covariance models

Statistical mechanics techniques have proved to be useful tools in quantifying the accuracy with which signal vectors are extracted from experimental data. However, analysis has previously been limited to specific model forms for the population covariance C, which may be inappropriate for real world data sets. In this paper we obtain new statistical mechanical results for a general population covariance matrix C. For data sets consisting of p sample points in we use the replica method to study the accuracy of orthogonal signal vectors estimated from the sample data. In the asymptotic limit of at fixed α = p/N, we derive analytical results for the signal direction learning curves. In the asymptotic limit the learning curves follow a single universal form, each displaying a retarded learning transition. An explicit formula for the location of the retarded learning transition is obtained and we find marked variation in the location of the retarded learning transition dependent on the distribution of population covariance eigenvalues. The results of the replica analysis are confirmed against simulation.

[1]  T. W. Anderson ASYMPTOTIC THEORY FOR PRINCIPAL COMPONENT ANALYSIS , 1963 .

[2]  I. Kanter Book Review: Statistical Mechanics of Learning. By A. Engel and C. Van den Broeck, Cambridge University Press , 2001 .

[3]  Statistical physics of independent component analysis , 2003, cond-mat/0309484.

[4]  Magnus Rattray,et al.  A Statistical Mechanics Analysis of Gram Matrix Eigenvalue Spectra , 2004, COLT.

[5]  S. Péché,et al.  Phase transition of the largest eigenvalue for nonnull complex sample covariance matrices , 2004, math/0403022.

[6]  M. Rattray,et al.  Statistical mechanics of learning multiple orthogonal signals: asymptotic theory and fluctuation effects. , 2007, Physical review. E, Statistical, nonlinear, and soft matter physics.

[7]  I. Johnstone On the distribution of the largest eigenvalue in principal components analysis , 2001 .

[8]  Z. Bai,et al.  METHODOLOGIES IN SPECTRAL ANALYSIS OF LARGE DIMENSIONAL RANDOM MATRICES, A REVIEW , 2008 .

[9]  P. R. Rider,et al.  Generalized cauchy distributions , 1957 .

[10]  V. Marčenko,et al.  DISTRIBUTION OF EIGENVALUES FOR SOME SETS OF RANDOM MATRICES , 1967 .

[11]  K. Wachter The Strong Limits of Random Matrix Spectra for Sample Matrices of Independent Elements , 1978 .

[12]  M. Rattray,et al.  Principal-component-analysis eigenvalue spectra from data with symmetry-breaking structure. , 2004, Physical review. E, Statistical, nonlinear, and soft matter physics.

[13]  Noureddine El Karoui Spectrum estimation for large dimensional covariance matrices using random matrix theory , 2006, math/0609418.

[14]  Michael Biehl An Exactly Solvable Model of Unsupervised Learning , 1994 .

[15]  John B. Thomas,et al.  Detectors for discrete-time signals in non-Gaussian noise , 1972, IEEE Trans. Inf. Theory.

[16]  Anirvan M. Sengupta,et al.  Distributions of singular values for some random matrices. , 1997, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[17]  D. Paul ASYMPTOTICS OF SAMPLE EIGENSTRUCTURE FOR A LARGE DIMENSIONAL SPIKED COVARIANCE MODEL , 2007 .

[18]  David C. Hoyle,et al.  Automatic PCA Dimension Selection for High Dimensional Data and Small Sample Sizes , 2008 .

[19]  J. W. Silverstein,et al.  Eigenvalues of large sample covariance matrices of spiked population models , 2004, math/0408165.

[20]  Magnus Rattray,et al.  PCA learning for sparse high-dimensional data , 2003 .

[21]  Geert Jan Bex,et al.  A Gaussian scenario for unsupervised learning , 1996 .