A-T-LAS$_{2,1}$: A Multi-Penalty Approach to Compressed Sensing of Low-Rank Matrices with Sparse Decompositions

Compressed Sensing is about recovering an unknown vector of dimension $n$ from $m \ll n$ linear measurements. This task becomes possible, for instance, when few entries of the vector have large magnitude and, hence, the vector is essentially of low intrinsic dimension. If one wishes to recover an $n_1\times n_2$ matrix instead, low-rankness can be added as sparsity structure of low intrinsic dimensionality. For this purpose we propose a novel algorithm, which we call Alternating Tikhonov regularization and Lasso (A-T-LAS$_{2,1}$). It is based on a multi-penalty regularization for recovery of low-rank matrices with approximately sparse singular vectors, which is able to leverage both structures (low-rankness and sparsity) simultaneously and effectively reduce further the number of necessary measurements with respect to the sole use of one of the two sparsity structures. We provide nearly-optimal recovery guarantees of A-T-LAS$_{2,1}$. The analysis is surprisingly relatively simple, e.g., compared to the one of other similar approaches such as Sparse Power Factorization (SPF). It relies on an adaptation of restricted isometry property to low-rank and approximately sparse matrices, LASSO technique, and results on proximal alternating minimization. We show that A-T-LAS$_{2,1}$ is more efficient than convex relaxation and exhibits similar performance to the state of the art method SPF, outperforming it in strong noise regimes and for matrices whose singular vectors do not possess exact (joint-) sparse support. Moreover, contrary to SPF, A-T-LAS$_{2,1}$, if properly initialized, is shown to converge even for measurements not fulfilling the restricted isometry property to matrices with some guaranteed sparsity and minimal discrepancy to data.

[1]  T. M. Cannon,et al.  Blind deconvolution through digital signal processing , 1975, Proceedings of the IEEE.

[2]  B. S. Kašin,et al.  DIAMETERS OF SOME FINITE-DIMENSIONAL SETS AND CLASSES OF SMOOTH FUNCTIONS , 1977 .

[3]  H. L. Taylor,et al.  Deconvolution with the l 1 norm , 1979 .

[4]  J R Fienup,et al.  Phase retrieval algorithms: a comparison. , 1982, Applied optics.

[5]  E. Gluskin NORMS OF RANDOM MATRICES AND WIDTHS OF FINITE-DIMENSIONAL SETS , 1984 .

[6]  F. Santosa,et al.  Linear inversion of ban limit reflection seismograms , 1986 .

[7]  B. Logan,et al.  Signal recovery and the large sieve , 1992 .

[8]  J. Drenth Principles of protein x-ray crystallography , 1994 .

[9]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[10]  Zhi Ding,et al.  Blind Equalization and Identification , 2001 .

[11]  I. Daubechies,et al.  An iterative thresholding algorithm for linear inverse problems with a sparsity constraint , 2003, math/0307152.

[12]  Joel A. Tropp,et al.  Greed is good: algorithmic results for sparse approximation , 2004, IEEE Transactions on Information Theory.

[13]  Liviu Badea,et al.  Sparse factorizations of gene expression data guided by binding data. , 2005, Pacific Symposium on Biocomputing. Pacific Symposium on Biocomputing.

[14]  E. Candès,et al.  Stable signal recovery from incomplete and inaccurate measurements , 2005, math/0503066.

[15]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[16]  David L Donoho,et al.  Compressed sensing , 2006, IEEE Transactions on Information Theory.

[17]  M. Rudelson,et al.  On sparse reconstruction from Fourier and Gaussian measurements , 2008 .

[18]  Mike E. Davies,et al.  Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.

[19]  Alternating minimization and projection methods for nonconvex problems , 2008 .

[20]  Pablo A. Parrilo,et al.  Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization , 2007, SIAM Rev..

[21]  Massimo Fornasier,et al.  Numerical Methods for Sparse Recovery , 2010 .

[22]  Deanna Needell,et al.  CoSaMP: Iterative signal recovery from incomplete and inaccurate samples , 2008, ArXiv.

[23]  Emmanuel J. Candès,et al.  Tight Oracle Inequalities for Low-Rank Matrix Recovery From a Minimal Number of Noisy Random Measurements , 2011, IEEE Transactions on Information Theory.

[24]  Emmanuel J. Candès,et al.  PhaseLift: Exact and Stable Signal Recovery from Magnitude Measurements via Convex Programming , 2011, ArXiv.

[25]  Yaniv Plan,et al.  One‐Bit Compressed Sensing by Linear Programming , 2011, ArXiv.

[26]  Roman Vershynin,et al.  Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.

[27]  Holger Rauhut,et al.  Suprema of Chaos Processes and the Restricted Isometry Property , 2012, ArXiv.

[28]  Sunav Choudhary,et al.  On identifiability in bilinear inverse problems , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[29]  Prateek Jain,et al.  Low-rank matrix completion using alternating minimization , 2012, STOC '13.

[30]  V. Naumova,et al.  Minimization of multi-penalty functionals by alternating iterative thresholding and optimal parameter choices , 2014, 1403.6718.

[31]  Justin K. Romberg,et al.  Blind Deconvolution Using Convex Programming , 2012, IEEE Transactions on Information Theory.

[32]  Yaniv Plan,et al.  Dimension Reduction by Random Hyperplane Tessellations , 2014, Discret. Comput. Geom..

[33]  Massimo Fornasier,et al.  Quasi-linear Compressed Sensing , 2013, Multiscale Model. Simul..

[34]  Justin K. Romberg,et al.  Near-Optimal Estimation of Simultaneously Sparse and Low-Rank Matrices from Nested Linear Measurements , 2015, ArXiv.

[35]  Yonina C. Eldar,et al.  Simultaneously Structured Models With Application to Sparse and Low-Rank Matrices , 2012, IEEE Transactions on Information Theory.

[36]  I. Daubechies,et al.  Sparsity-enforcing regularisation and ISTA revisited , 2016 .

[37]  M. Grasmair,et al.  Conditions on optimal support recovery in unmixing problems by means of multi-penalty regularization , 2016, 1601.01461.

[38]  Timo Klock,et al.  Adaptive multi-penalty regularization based on a generalized Lasso path , 2017, Applied and Computational Harmonic Analysis.

[39]  F. Krahmer,et al.  Refined performance guarantees for Sparse Power Factorization , 2017, 2017 International Conference on Sampling Theory and Applications (SampTA).

[40]  Yoram Bresler,et al.  Near-Optimal Compressed Sensing of a Class of Sparse Low-Rank Matrices Via Sparse Power Factorization , 2013, IEEE Transactions on Information Theory.

[41]  Peter Jung,et al.  Blind Demixing and Deconvolution at Near-Optimal Rate , 2017, IEEE Transactions on Information Theory.