On the Theorem of Uniform Recovery of Random Sampling Matrices

We consider two theorems from the theory of compressive sensing. Mainly a theorem concerning uniform recovery of random sampling matrices, where the number of samples needed in order to recover an s-sparse signal from linear measurements (with high probability) is known to be m ≳ s(ln s)<;sup>3<;/sup>lnN. We present new and improved constants together with what we consider to be a more explicit proof. A proof that also allows for a slightly larger class of m × N-matrices, by considering what is called effective sparsity. We also present a condition on the so-called restricted isometry constants, δ<;sub>s<;/sub>, ensuring sparse recovery via ℓ<;sup>1<;/sup>-minimization. We show that is sufficient and that this can be improved further to almost allow for a sufficient condition of the type .

[1]  H. Rauhut Compressive Sensing and Structured Random Matrices , 2009 .

[2]  Uffe Haagerup,et al.  On the best constants in noncommutative Khintchine-type inequalities , 2006, math/0611160.

[3]  Emmanuel J. Candès,et al.  A Probabilistic and RIPless Theory of Compressed Sensing , 2010, IEEE Transactions on Information Theory.

[4]  Massimo Fornasier,et al.  Compressive Sensing and Structured Random Matrices , 2010 .

[5]  Lie Wang,et al.  New Bounds for Restricted Isometry Constants , 2009, IEEE Transactions on Information Theory.

[6]  Venkatesan Guruswami,et al.  Restricted Isometry of Fourier Matrices and List Decodability of Random Linear Codes , 2012, SIAM J. Comput..

[7]  Yaniv Plan,et al.  One‐Bit Compressed Sensing by Linear Programming , 2011, ArXiv.

[8]  J. Kuelbs Probability on Banach spaces , 1978 .

[9]  S. Foucart A note on guaranteed sparse recovery via ℓ1-minimization , 2010 .

[10]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[11]  Massimo Fornasier,et al.  Theoretical Foundations and Numerical Methods for Sparse Recovery , 2010, Radon Series on Computational and Applied Mathematics.

[12]  H. Rauhut,et al.  Interpolation via weighted $l_1$ minimization , 2013, 1308.0759.

[13]  Rémi Gribonval,et al.  Restricted Isometry Constants Where $\ell ^{p}$ Sparse Recovery Can Fail for $0≪ p \leq 1$ , 2009, IEEE Transactions on Information Theory.

[14]  慧 廣瀬 A Mathematical Introduction to Compressive Sensing , 2015 .

[15]  Anru Zhang,et al.  Sparse Representation of a Polytope and Recovery of Sparse Signals and Low-Rank Matrices , 2013, IEEE Transactions on Information Theory.

[16]  Davies Rémi Gribonval Restricted Isometry Constants Where Lp Sparse Recovery Can Fail for 0 , 2008 .

[17]  M. Rudelson,et al.  On sparse reconstruction from Fourier and Gaussian measurements , 2008 .

[18]  S. Frick,et al.  Compressed Sensing , 2014, Computer Vision, A Reference Guide.

[19]  Axthonv G. Oettinger,et al.  IEEE Transactions on Information Theory , 1998 .

[20]  Gongguo Tang,et al.  Performance Analysis of Sparse Recovery Based on Constrained Minimal Singular Values , 2010, IEEE Transactions on Signal Processing.

[21]  Pär Kurlberg,et al.  A local Riemann hypothesis, I , 2000 .

[22]  Song Li,et al.  New bounds on the restricted isometry constant δ2k , 2011 .

[23]  Rémi Gribonval,et al.  Sparse representations in unions of bases , 2003, IEEE Trans. Inf. Theory.

[24]  J. Wissel,et al.  On the Best Constants in the Khintchine Inequality , 2007 .

[25]  Lie Wang,et al.  Shifting Inequality and Recovery of Sparse Signals , 2010, IEEE Transactions on Signal Processing.

[26]  M. Lifshits Lectures on Gaussian Processes , 2012 .

[27]  Emmanuel J. Candès,et al.  Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.