Online Low-Rank Subspace Learning from Incomplete Data: A Bayesian View

Extracting the underlying low-dimensional space where high-dimensional signals often reside has long been at the center of numerous algorithms in the signal processing and machine learning literature during the past few decades. At the same time, working with incomplete (partly observed) large scale datasets has recently been commonplace for diverse reasons. This so called {\it big data era} we are currently living calls for devising online subspace learning algorithms that can suitably handle incomplete data. Their envisaged objective is to {\it recursively} estimate the unknown subspace by processing streaming data sequentially, thus reducing computational complexity, while obviating the need for storing the whole dataset in memory. In this paper, an online variational Bayes subspace learning algorithm from partial observations is presented. To account for the unawareness of the true rank of the subspace, commonly met in practice, low-rankness is explicitly imposed on the sought subspace data matrix by exploiting sparse Bayesian learning principles. Moreover, sparsity, {\it simultaneously} to low-rankness, is favored on the subspace matrix by the sophisticated hierarchical Bayesian scheme that is adopted. In doing so, the proposed algorithm becomes adept in dealing with applications whereby the underlying subspace may be also sparse, as, e.g., in sparse dictionary learning problems. As shown, the new subspace tracking scheme outperforms its state-of-the-art counterparts in terms of estimation accuracy, in a variety of experiments conducted on simulated and real data.

[1]  C. Févotte,et al.  Automatic Relevance Determination in Nonnegative Matrix Factorization , 2009 .

[2]  Robert D. Nowak,et al.  Online identification and tracking of subspaces from highly incomplete information , 2010, 2010 48th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[3]  J. Bunch,et al.  Rank-one modification of the symmetric eigenproblem , 1978 .

[4]  Tom Minka,et al.  Expectation Propagation for approximate Bayesian inference , 2001, UAI.

[5]  George Eastman House,et al.  Sparse Bayesian Learning and the Relevan e Ve tor Ma hine , 2001 .

[6]  S. Gull,et al.  Image reconstruction from incomplete and noisy data , 1978, Nature.

[7]  Konstantinos Koutroumbas,et al.  Online Bayesian low-rank subspace learning from partial observations , 2015, 2015 23rd European Signal Processing Conference (EUSIPCO).

[8]  Guillermo Sapiro,et al.  Online Learning for Matrix Factorization and Sparse Coding , 2009, J. Mach. Learn. Res..

[9]  Bin Yang,et al.  Projection approximation subspace tracking , 1995, IEEE Trans. Signal Process..

[10]  George V. Moustakides,et al.  Fast and Stable Subspace Tracking , 2008, IEEE Transactions on Signal Processing.

[11]  Laurent El Ghaoui,et al.  Large-Scale Sparse Principal Component Analysis with Application to Text Data , 2011, NIPS.

[12]  A. Taylan Cemgil,et al.  Chapter 19 - A Tutorial Introduction to Monte Carlo Methods, Markov Chain Monte Carlo and Particle Filtering , 2014 .

[13]  Morteza Mardani,et al.  Dynamic Anomalography: Tracking Network Anomalies Via Sparsity and Low Rank , 2012, IEEE Journal of Selected Topics in Signal Processing.

[14]  Taghi M. Khoshgoftaar,et al.  A Survey of Collaborative Filtering Techniques , 2009, Adv. Artif. Intell..

[15]  Masa-aki Sato,et al.  Online Model Selection Based on the Variational Bayes , 2001, Neural Computation.

[16]  Lawrence Carin,et al.  Active learning for online bayesian matrix factorization , 2012, KDD.

[17]  Konstantinos Koutroumbas,et al.  A Variational Bayes Framework for Sparse Adaptive Estimation , 2014, IEEE Transactions on Signal Processing.

[18]  Norman L. Owsley,et al.  Adaptive data orthogonalization , 1978, ICASSP.

[19]  Huan Xu,et al.  Streaming Sparse Principal Component Analysis , 2015, ICML.

[20]  Dimitri P. Bertsekas,et al.  Nonlinear Programming , 1997 .

[21]  Emmanuel J. Candès,et al.  Exact Matrix Completion via Convex Optimization , 2009, Found. Comput. Math..

[22]  Adi Shraibman,et al.  Rank, Trace-Norm and Max-Norm , 2005, COLT.

[23]  Paris V. Giampouras,et al.  Simultaneously Sparse and Low-Rank Abundance Matrix Estimation for Hyperspectral Image Unmixing , 2015, IEEE Transactions on Geoscience and Remote Sensing.

[24]  Sergios Theodoridis,et al.  Machine Learning: A Bayesian and Optimization Perspective , 2015 .

[25]  Aggelos K. Katsaggelos,et al.  Sparse Bayesian Methods for Low-Rank Matrix Estimation , 2011, IEEE Transactions on Signal Processing.

[26]  Judith Rousseau,et al.  Bayesian matrix completion: prior specification and consistency , 2014 .

[27]  A. Robert Calderbank,et al.  PETRELS: Parallel Subspace Estimation and Tracking by Recursive Least Squares From Partial Observations , 2012, IEEE Transactions on Signal Processing.

[28]  Stephen J. Wright,et al.  Local Convergence of an Algorithm for Subspace Identification from Partial Data , 2013, Found. Comput. Math..

[29]  Jiawei Han,et al.  Spectral Regression: A Unified Approach for Sparse Subspace Learning , 2007, Seventh IEEE International Conference on Data Mining (ICDM 2007).

[30]  Morteza Mardani,et al.  Subspace Learning and Imputation for Streaming Big Data Matrices and Tensors , 2014, IEEE Transactions on Signal Processing.

[31]  M. Kowalski Sparse regression using mixed norms , 2009 .

[32]  Michael Elad,et al.  Image Denoising Via Sparse and Redundant Representations Over Learned Dictionaries , 2006, IEEE Transactions on Image Processing.

[33]  D.G. Tzikas,et al.  The variational approximation for Bayesian inference , 2008, IEEE Signal Processing Magazine.