One-Bit Compressive Sensing With Projected Subgradient Method Under Sparsity Constraints

One-bit compressive sensing theory shows that the sparse signals can be almost exactly reconstructed from a small number of one-bit quantized linear measurements. This paper presents the convergence analysis of the binary iterative hard thresholding (BIHT) algorithm which is a state-of-the-art recovery algorithm in one-bit compressive sensing. The basic idea of the convergence analysis is to view BIHT as a kind of projected subgradient method under sparsity constrains. To the best of our knowledge, this is the first convergence analysis of BIHT. We first consider a general convex function subject to sparsity constraints and connect it with the non-convex model in one-bit compressive sensing literatures. A projected subgradient method is proposed to solve the general model and some convergence results are established. A stronger convergence theorem for $\alpha $ –strongly convex functions without assumption on differentiable condition is also established. Furthermore, the corresponding stochastic projected subgradient method is provided with convergence guarantee. In our settings, BIHT is a special case of the projected subgradient method. Therefore, the convergence analysis can be applied to BIHT naturally. Then, we apply the projected subgradient method to some related non-convex optimization models arising in compressive sensing with $\ell _{1}$ -constraint, sparse support vector machines, and rectifier linear units regression. Finally, some numerical examples are presented to show the validity of our convergence analysis. The numerical experiments also show that the proposed projected subgradient method is very simple to implement, robust to sparse noise, and effective for sparse recovery problems.

[1]  Mark W. Schmidt,et al.  A simpler approach to obtaining an O(1/t) convergence rate for the projected stochastic subgradient method , 2012, ArXiv.

[2]  Gitta Kutyniok,et al.  1 . 2 Sparsity : A Reasonable Assumption ? , 2012 .

[3]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[4]  Sungyoung Lee,et al.  Compressive sensing: From theory to applications, a survey , 2013, Journal of Communications and Networks.

[5]  Mike E. Davies,et al.  Sampling Theorems for Signals From the Union of Finite-Dimensional Linear Subspaces , 2009, IEEE Transactions on Information Theory.

[6]  Alfredo N. Iusem,et al.  On the projected subgradient method for nonsmooth convex optimization in a Hilbert space , 1998, Math. Program..

[7]  Deanna Needell,et al.  Linear Convergence of Stochastic Iterative Greedy Algorithms With Sparse Constraints , 2014, IEEE Transactions on Information Theory.

[8]  Vikas Sindhwani,et al.  Efficient and Practical Stochastic Subgradient Descent for Nuclear Norm Regularization , 2012, ICML.

[9]  Yurii Nesterov,et al.  Primal-dual subgradient methods for convex problems , 2005, Math. Program..

[10]  Ohad Shamir,et al.  Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.

[11]  Jared Tanner,et al.  Normalized Iterative Hard Thresholding for Matrix Completion , 2013, SIAM J. Sci. Comput..

[12]  H. Robbins A Stochastic Approximation Method , 1951 .

[13]  Emmanuel J. Candès,et al.  Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.

[14]  E. Candès Mathematics of Sparsity (and a Few Other Things) , 2014 .

[15]  Marc Teboulle,et al.  Mirror descent and nonlinear projected subgradient methods for convex optimization , 2003, Oper. Res. Lett..

[16]  Stephen P. Boyd,et al.  Stochastic Subgradient Methods , 2007 .

[17]  Elena Braverman,et al.  Stable recovery of analysis based approaches , 2015 .

[18]  Yaniv Plan,et al.  One‐Bit Compressed Sensing by Linear Programming , 2011, ArXiv.

[19]  Deanna Needell,et al.  Methods for quantized compressed sensing , 2015, 2016 Information Theory and Applications Workshop (ITA).

[20]  Z. Bai,et al.  Limit of the smallest eigenvalue of a large dimensional sample covariance matrix , 1993 .

[21]  Ming Yan,et al.  Robust 1-bit Compressive Sensing Using Adaptive Outlier Pursuit , 2012, IEEE Transactions on Signal Processing.

[22]  T. Blumensath,et al.  Iterative Thresholding for Sparse Approximations , 2008 .

[23]  Yi Shen,et al.  The proximal alternating iterative hard thresholding method for l0 minimization, with complexity , 2017, J. Comput. Appl. Math..

[24]  Wotao Yin,et al.  Trust, But Verify: Fast and Accurate Signal Recovery From 1-Bit Compressive Measurements , 2011, IEEE Transactions on Signal Processing.

[25]  Z. Bai,et al.  On the limit of the largest eigenvalue of the large dimensional sample covariance matrix , 1988 .

[26]  R. Barber,et al.  Gradient descent with nonconvex constraints: local concavity determines convergence , 2017, 1703.07755.

[27]  Yuesheng Xu,et al.  Wavelet inpainting with the ℓ0 sparse regularization , 2016 .

[28]  Song Li,et al.  Convergence analysis of projected gradient descent for Schatten-p nonconvex matrix recovery , 2015 .

[29]  Bernhard Schölkopf,et al.  Use of the Zero-Norm with Linear Models and Kernel Methods , 2003, J. Mach. Learn. Res..

[30]  Deanna Needell,et al.  CoSaMP: Iterative signal recovery from incomplete and inaccurate samples , 2008, ArXiv.

[31]  Emmanuel J. Candès,et al.  Decoding by linear programming , 2005, IEEE Transactions on Information Theory.

[32]  Martin J. Wainwright,et al.  Fast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees , 2015, ArXiv.

[33]  Klaus-Robert Müller,et al.  Efficient BackProp , 2012, Neural Networks: Tricks of the Trade.

[34]  Olgica Milenkovic,et al.  Subspace Pursuit for Compressive Sensing Signal Reconstruction , 2008, IEEE Transactions on Information Theory.

[35]  Laurent Jacques,et al.  Robust 1-Bit Compressive Sensing via Binary Stable Embeddings of Sparse Vectors , 2011, IEEE Transactions on Information Theory.

[36]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[37]  Mike E. Davies,et al.  Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.

[38]  Laurent Jacques,et al.  Quantized Iterative Hard Thresholding: Bridging 1-bit and High-Resolution Quantized Compressed Sensing , 2013, ArXiv.

[39]  P. Maingé Strong Convergence of Projected Subgradient Methods for Nonsmooth and Nonstrictly Convex Minimization , 2008 .

[40]  David L Donoho,et al.  Compressed sensing , 2006, IEEE Transactions on Information Theory.

[41]  Nadav Hallak,et al.  On the Minimization Over Sparse Symmetric Sets: Projections, Optimality Conditions, and Algorithms , 2016, Math. Oper. Res..

[42]  Zhaosong Lu,et al.  Iterative hard thresholding methods for l0\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$l_0$$\end{document} regulari , 2012, Mathematical Programming.

[43]  Stephen P. Boyd,et al.  Subgradient Methods , 2007 .

[44]  Yuesheng Xu,et al.  Noisy 1-bit compressive sensing: Models and algorithms , 2016 .

[45]  Lie Wang The L1L1 penalized LAD estimator for high dimensional linear regression , 2013, J. Multivar. Anal..

[46]  Thomas Blumensath,et al.  Compressed Sensing With Nonlinear Observations and Related Nonlinear Optimization Problems , 2012, IEEE Transactions on Information Theory.

[47]  Yonina C. Eldar,et al.  Robust Recovery of Signals From a Structured Union of Subspaces , 2008, IEEE Transactions on Information Theory.

[48]  Zhiqiang Xu,et al.  One-Bit Compressed Sensing by Greedy Algorithms , 2013, ArXiv.

[49]  Minh N. Do,et al.  A Theory for Sampling Signals from a Union of Subspaces , 2022 .

[50]  Junhong Lin,et al.  Convergence of projected Landweber iteration for matrix rank minimization , 2014 .

[51]  Roman Vershynin,et al.  Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.

[52]  Yonina C. Eldar,et al.  Compressed Sensing with Coherent and Redundant Dictionaries , 2010, ArXiv.

[53]  Lixin Shen,et al.  One-bit compressive sampling via ℓ0 minimization , 2016, EURASIP J. Adv. Signal Process..

[54]  Ohad Shamir,et al.  Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes , 2012, ICML.

[55]  Song Li,et al.  Sparse Recovery with Coherent Tight Frame via Analysis Dantzig Selector and Analysis LASSO , 2013, ArXiv.

[56]  Mahdi Soltanolkotabi,et al.  Learning ReLUs via Gradient Descent , 2017, NIPS.

[57]  Po-Ling Loh,et al.  Regularized M-estimators with nonconvexity: statistical and algorithmic theory for local optima , 2013, J. Mach. Learn. Res..

[58]  Yonina C. Eldar,et al.  Sparsity Constrained Nonlinear Optimization: Optimality Conditions and Algorithms , 2012, SIAM J. Optim..

[59]  Richard G. Baraniuk,et al.  1-Bit compressive sensing , 2008, 2008 42nd Annual Conference on Information Sciences and Systems.

[60]  Sébastien Bubeck,et al.  Convex Optimization: Algorithms and Complexity , 2014, Found. Trends Mach. Learn..