One-Bit Compressive Sensing With Projected Subgradient Method Under Sparsity Constraints
暂无分享,去创建一个
[1] Mark W. Schmidt,et al. A simpler approach to obtaining an O(1/t) convergence rate for the projected stochastic subgradient method , 2012, ArXiv.
[2] Gitta Kutyniok,et al. 1 . 2 Sparsity : A Reasonable Assumption ? , 2012 .
[3] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[4] Sungyoung Lee,et al. Compressive sensing: From theory to applications, a survey , 2013, Journal of Communications and Networks.
[5] Mike E. Davies,et al. Sampling Theorems for Signals From the Union of Finite-Dimensional Linear Subspaces , 2009, IEEE Transactions on Information Theory.
[6] Alfredo N. Iusem,et al. On the projected subgradient method for nonsmooth convex optimization in a Hilbert space , 1998, Math. Program..
[7] Deanna Needell,et al. Linear Convergence of Stochastic Iterative Greedy Algorithms With Sparse Constraints , 2014, IEEE Transactions on Information Theory.
[8] Vikas Sindhwani,et al. Efficient and Practical Stochastic Subgradient Descent for Nuclear Norm Regularization , 2012, ICML.
[9] Yurii Nesterov,et al. Primal-dual subgradient methods for convex problems , 2005, Math. Program..
[10] Ohad Shamir,et al. Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.
[11] Jared Tanner,et al. Normalized Iterative Hard Thresholding for Matrix Completion , 2013, SIAM J. Sci. Comput..
[12] H. Robbins. A Stochastic Approximation Method , 1951 .
[13] Emmanuel J. Candès,et al. Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information , 2004, IEEE Transactions on Information Theory.
[14] E. Candès. Mathematics of Sparsity (and a Few Other Things) , 2014 .
[15] Marc Teboulle,et al. Mirror descent and nonlinear projected subgradient methods for convex optimization , 2003, Oper. Res. Lett..
[16] Stephen P. Boyd,et al. Stochastic Subgradient Methods , 2007 .
[17] Elena Braverman,et al. Stable recovery of analysis based approaches , 2015 .
[18] Yaniv Plan,et al. One‐Bit Compressed Sensing by Linear Programming , 2011, ArXiv.
[19] Deanna Needell,et al. Methods for quantized compressed sensing , 2015, 2016 Information Theory and Applications Workshop (ITA).
[20] Z. Bai,et al. Limit of the smallest eigenvalue of a large dimensional sample covariance matrix , 1993 .
[21] Ming Yan,et al. Robust 1-bit Compressive Sensing Using Adaptive Outlier Pursuit , 2012, IEEE Transactions on Signal Processing.
[22] T. Blumensath,et al. Iterative Thresholding for Sparse Approximations , 2008 .
[23] Yi Shen,et al. The proximal alternating iterative hard thresholding method for l0 minimization, with complexity , 2017, J. Comput. Appl. Math..
[24] Wotao Yin,et al. Trust, But Verify: Fast and Accurate Signal Recovery From 1-Bit Compressive Measurements , 2011, IEEE Transactions on Signal Processing.
[25] Z. Bai,et al. On the limit of the largest eigenvalue of the large dimensional sample covariance matrix , 1988 .
[26] R. Barber,et al. Gradient descent with nonconvex constraints: local concavity determines convergence , 2017, 1703.07755.
[27] Yuesheng Xu,et al. Wavelet inpainting with the ℓ0 sparse regularization , 2016 .
[28] Song Li,et al. Convergence analysis of projected gradient descent for Schatten-p nonconvex matrix recovery , 2015 .
[29] Bernhard Schölkopf,et al. Use of the Zero-Norm with Linear Models and Kernel Methods , 2003, J. Mach. Learn. Res..
[30] Deanna Needell,et al. CoSaMP: Iterative signal recovery from incomplete and inaccurate samples , 2008, ArXiv.
[31] Emmanuel J. Candès,et al. Decoding by linear programming , 2005, IEEE Transactions on Information Theory.
[32] Martin J. Wainwright,et al. Fast low-rank estimation by projected gradient descent: General statistical and algorithmic guarantees , 2015, ArXiv.
[33] Klaus-Robert Müller,et al. Efficient BackProp , 2012, Neural Networks: Tricks of the Trade.
[34] Olgica Milenkovic,et al. Subspace Pursuit for Compressive Sensing Signal Reconstruction , 2008, IEEE Transactions on Information Theory.
[35] Laurent Jacques,et al. Robust 1-Bit Compressive Sensing via Binary Stable Embeddings of Sparse Vectors , 2011, IEEE Transactions on Information Theory.
[36] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[37] Mike E. Davies,et al. Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.
[38] Laurent Jacques,et al. Quantized Iterative Hard Thresholding: Bridging 1-bit and High-Resolution Quantized Compressed Sensing , 2013, ArXiv.
[39] P. Maingé. Strong Convergence of Projected Subgradient Methods for Nonsmooth and Nonstrictly Convex Minimization , 2008 .
[40] David L Donoho,et al. Compressed sensing , 2006, IEEE Transactions on Information Theory.
[41] Nadav Hallak,et al. On the Minimization Over Sparse Symmetric Sets: Projections, Optimality Conditions, and Algorithms , 2016, Math. Oper. Res..
[42] Zhaosong Lu,et al. Iterative hard thresholding methods for l0\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$l_0$$\end{document} regulari , 2012, Mathematical Programming.
[43] Stephen P. Boyd,et al. Subgradient Methods , 2007 .
[44] Yuesheng Xu,et al. Noisy 1-bit compressive sensing: Models and algorithms , 2016 .
[45] Lie Wang. The L1L1 penalized LAD estimator for high dimensional linear regression , 2013, J. Multivar. Anal..
[46] Thomas Blumensath,et al. Compressed Sensing With Nonlinear Observations and Related Nonlinear Optimization Problems , 2012, IEEE Transactions on Information Theory.
[47] Yonina C. Eldar,et al. Robust Recovery of Signals From a Structured Union of Subspaces , 2008, IEEE Transactions on Information Theory.
[48] Zhiqiang Xu,et al. One-Bit Compressed Sensing by Greedy Algorithms , 2013, ArXiv.
[49] Minh N. Do,et al. A Theory for Sampling Signals from a Union of Subspaces , 2022 .
[50] Junhong Lin,et al. Convergence of projected Landweber iteration for matrix rank minimization , 2014 .
[51] Roman Vershynin,et al. Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.
[52] Yonina C. Eldar,et al. Compressed Sensing with Coherent and Redundant Dictionaries , 2010, ArXiv.
[53] Lixin Shen,et al. One-bit compressive sampling via ℓ0 minimization , 2016, EURASIP J. Adv. Signal Process..
[54] Ohad Shamir,et al. Stochastic Gradient Descent for Non-smooth Optimization: Convergence Results and Optimal Averaging Schemes , 2012, ICML.
[55] Song Li,et al. Sparse Recovery with Coherent Tight Frame via Analysis Dantzig Selector and Analysis LASSO , 2013, ArXiv.
[56] Mahdi Soltanolkotabi,et al. Learning ReLUs via Gradient Descent , 2017, NIPS.
[57] Po-Ling Loh,et al. Regularized M-estimators with nonconvexity: statistical and algorithmic theory for local optima , 2013, J. Mach. Learn. Res..
[58] Yonina C. Eldar,et al. Sparsity Constrained Nonlinear Optimization: Optimality Conditions and Algorithms , 2012, SIAM J. Optim..
[59] Richard G. Baraniuk,et al. 1-Bit compressive sensing , 2008, 2008 42nd Annual Conference on Information Sciences and Systems.
[60] Sébastien Bubeck,et al. Convex Optimization: Algorithms and Complexity , 2014, Found. Trends Mach. Learn..