Testing Sparsity over Known and Unknown Bases

Sparsity is a basic property of real vectors that is exploited in a wide variety of applications. In this work, we describe property testing algorithms for sparsity that observe a low-dimensional projection of the input. We consider two settings. In the first setting, for a given design matrix A in R^{d x m}, we test whether an input vector y in R^d equals Ax for some k-sparse unit vector x. Our algorithm projects the input onto O(k \eps^{-2} log m) dimensions, accepts if the property holds, rejects if ||y - Ax|| > \eps for any O(k/\eps^2)-sparse vector x, and runs in time polynomial in m. Our algorithm is based on the approximate Caratheodory's theorem. Previously known algorithms that solve the problem for arbitrary A with qualitatively similar guarantees run in exponential time. In the second setting, the design matrix A is unknown. Given input vectors y_1, y_2,...,y_p in R^d whose concatenation as columns forms Y in R^{d x p} , the goal is to decide whether Y=AX for matrices A in R^{d x m} and X in R^{m x p} such that each column of X is k-sparse, or whether Y is "far" from having such a decomposition. We give such a testing algorithm which projects the input vectors to O(log p/\eps^2) dimensions and assumes that the unknown A satisfies k-restricted isometry. Our analysis gives a new robust characterization of gaussian width in terms of sparsity.

[1]  Y. Gordon Some inequalities for Gaussian processes and applications , 1985 .

[2]  M. Rudelson,et al.  On sparse reconstruction from Fourier and Gaussian measurements , 2008 .

[3]  Sjoerd Dirksen,et al.  Toward a unified theory of sparse dimensionality reduction in Euclidean space , 2013, STOC.

[4]  Dana Ron Property Testing: A Learning Theory Perspective , 2008, Found. Trends Mach. Learn..

[5]  Ronitt Rubinfeld,et al.  Sublinear Time Algorithms , 2011, SIAM J. Discret. Math..

[6]  Noga Alon,et al.  The Space Complexity of Approximating the Frequency Moments , 1999 .

[7]  Michael Elad,et al.  Image Denoising Via Sparse and Redundant Representations Over Learned Dictionaries , 2006, IEEE Transactions on Image Processing.

[8]  Prateek Jain,et al.  Learning Sparsely Used Overcomplete Dictionaries , 2014, COLT.

[9]  Noga Alon,et al.  The approximate rank of a matrix and its algorithmic applications: approximate rank , 2013, STOC '13.

[10]  R. Vershynin Lectures in Geometric Functional Analysis , 2012 .

[11]  D. Slepian The one-sided barrier problem for Gaussian noise , 1962 .

[12]  R. Vershynin Estimation in High Dimensions: A Geometric Perspective , 2014, 1405.5103.

[13]  Dana Ron,et al.  Property testing and its connection to learning and approximation , 1998, JACM.

[14]  David P. Woodruff Sketching as a Tool for Numerical Linear Algebra , 2014, Found. Trends Theor. Comput. Sci..

[15]  Huan Wang,et al.  Exact Recovery of Sparsely-Used Dictionaries , 2012, COLT.

[16]  C. O’Brien Statistical Learning with Sparsity: The Lasso and Generalizations , 2016 .

[17]  Andrew McGregor,et al.  Graph stream algorithms: a survey , 2014, SGMD.

[18]  Dean P. Foster,et al.  Variable Selection is Hard , 2014, COLT.

[19]  Silvio Lattanzi,et al.  Algorithms for $\ell_p$ Low-Rank Approximation , 2017, ICML.

[20]  David P. Woodruff,et al.  An optimal algorithm for the distinct elements problem , 2010, PODS '10.

[21]  Carsten Lund,et al.  Non-deterministic exponential time has two-prover interactive protocols , 2005, computational complexity.

[22]  Joel A. Tropp,et al.  Living on the edge: A geometric theory of phase transitions in convex optimization , 2013, ArXiv.

[23]  Artur Czumaj,et al.  Property Testing in Computational Geometry , 2000, ESA.

[24]  Peter L. Bartlett,et al.  Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..

[25]  Manuel Blum,et al.  Self-Testing/Correcting with Applications to Numerical Problems , 1993, J. Comput. Syst. Sci..

[26]  Anirban Dasgupta,et al.  A sparse Johnson: Lindenstrauss transform , 2010, STOC '10.

[27]  Shahar Mendelson,et al.  Entropy, Combinatorial Dimensions and Random Averages , 2002, COLT.

[28]  Sariel Har-Peled,et al.  Sparse Approximation via Generating Point Sets , 2015, SODA.

[29]  Zhi-Quan Luo,et al.  Dictionary learning for sparse representation: Complexity and algorithms , 2014, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[30]  Balas K. Natarajan,et al.  Sparse Approximate Solutions to Linear Systems , 1995, SIAM J. Comput..

[31]  Gábor Lugosi,et al.  Concentration Inequalities - A Nonasymptotic Theory of Independence , 2013, Concentration Inequalities.

[32]  Pablo A. Parrilo,et al.  The Convex Geometry of Linear Inverse Problems , 2010, Foundations of Computational Mathematics.

[33]  Ronitt Rubinfeld,et al.  Robust Characterizations of Polynomials with Applications to Program Testing , 1996, SIAM J. Comput..

[34]  Ori Sasson,et al.  Property testing of data dimensionality , 2003, SODA '03.

[35]  Daniel M. Kane,et al.  Sparser Johnson-Lindenstrauss Transforms , 2010, JACM.

[36]  Andreas M. Tillmann On the Computational Intractability of Exact and Approximate Dictionary Learning , 2014, IEEE Signal Processing Letters.