Bayesian Dimensionality Reduction With PCA Using Penalized Semi-Integrated Likelihood

ABSTRACT We discuss the problem of estimating the number of principal components in principal components analysis (PCA). Despite the importance of the problem and the multitude of solutions proposed in literature, it comes as a surprise that there does not exist a coherent asymptotic framework, which would justify different approaches depending on the actual size of the dataset. In this article, we address this issue by presenting an approximate Bayesian approach based on Laplace approximation and introducing a general method of developing criteria for model selection, called PEnalized SEmi-integrated Likelihood (PESEL). Our general framework encompasses a variety of existing approaches based on probabilistic models, like the Bayesian Information Criterion for Probabilistic PCA (PPCA), and enables the construction of new criteria, depending on the size of the dataset at hand and additional prior information. Specifically, we apply PESEL to derive two new criteria for datasets where the number of variables substantially exceeds the number of observations, which is out of the scope of currently existing approaches. We also report results of extensive simulation studies and real data analysis, which illustrate the desirable properties of our proposed criteria as compared to state-of-the-art methods and very recent proposals. Specifically, these simulations show that PESEL-based criteria can be quite robust against deviations from the assumptions of a probabilistic model. Selected PESEL-based criteria for the estimation of the number of principal components are implemented in the R package pesel, which is available on github (https://github.com/psobczyk/pesel). Supplementary material for this article, with additional simulation results, is available online. The code to reproduce all simulations is available at https://github.com/psobczyk/pesel_simulations.

[1]  Michael E. Tipping,et al.  Probabilistic Principal Component Analysis , 1999 .

[2]  J. Pagès,et al.  Gestion des données manquantes en analyse en composantes principales , 2009 .

[3]  J. Bai,et al.  Determining the Number of Factors in Approximate Factor Models , 2000 .

[4]  Julie Josse,et al.  Selecting the number of components in principal component analysis using cross-validation approximations , 2012, Comput. Stat. Data Anal..

[5]  Tapani Raiko,et al.  Tkk Reports in Information and Computer Science Practical Approaches to Principal Component Analysis in the Presence of Missing Values Tkk Reports in Information and Computer Science Practical Approaches to Principal Component Analysis in the Presence of Missing Values , 2022 .

[6]  Christopher M. Bishop,et al.  Bayesian PCA , 1998, NIPS.

[7]  Emmanuel J. Candès,et al.  Randomized Algorithms for Low-Rank Matrix Factorizations: Sharp Performance Bounds , 2013, Algorithmica.

[8]  R Core Team,et al.  R: A language and environment for statistical computing. , 2014 .

[9]  Y. Chikuse Statistics on special manifolds , 2003 .

[10]  B. Ninness Estimation of Noise , 1998 .

[11]  David C. Hoyle,et al.  Automatic PCA Dimension Selection for High Dimensional Data and Small Sample Sizes , 2008 .

[12]  Karl Pearson F.R.S. LIII. On lines and planes of closest fit to systems of points in space , 1901 .

[13]  Peter D. Hoff,et al.  Model Averaging and Dimension Selection for the Singular Value Decomposition , 2006, math/0609042.

[14]  Patrick O. Perry,et al.  Bi-cross-validation of the SVD and the nonnegative matrix factorization , 2009, 0908.2062.

[15]  Damien Passemier,et al.  On estimation of the noise variance in high dimensional probabilistic principal component analysis , 2013, 1308.3890.

[16]  Nabil H. Mustafa,et al.  k-means projective clustering , 2004, PODS.

[17]  Jérôme Pagès,et al.  Multiple imputation in principal component analysis , 2011, Adv. Data Anal. Classif..

[18]  Tom Minka,et al.  Automatic Choice of Dimensionality for PCA , 2000, NIPS.

[19]  N. D. Bruijn Asymptotic methods in analysis , 1958 .

[20]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[21]  J. J. Rajan,et al.  Model Order Selection For The Singular Value Decomposition And The Discrete Karhunen-Loeve Transform Using A Bayesian Approach , 1997 .

[22]  A. James Normal Multivariate Analysis and the Orthogonal Group , 1954 .

[23]  I. Jolliffe Principal Component Analysis , 2002 .

[24]  Genevera I. Allen,et al.  Journal of the American Statistical Association a Generalized Least-square Matrix Decomposition a Generalized Least-square Matrix Decomposition , 2022 .

[25]  Gary Evans,et al.  Exploratory Multivariate Analysis by Example Using R , 2011 .

[26]  Shinichi Nakajima,et al.  Condition for perfect dimensionality recovery by variational Bayesian PCA , 2015, J. Mach. Learn. Res..

[27]  Donald A. Jackson STOPPING RULES IN PRINCIPAL COMPONENTS ANALYSIS: A COMPARISON OF HEURISTICAL AND STATISTICAL APPROACHES' , 1993 .

[28]  R. Tibshirani,et al.  Selecting the number of principal components: estimation of the true rank of a noisy matrix , 2014, 1410.8260.

[29]  Charles M. Bishop Variational principal components , 1999 .