A Constructive Approach to $L_0$ Penalized Regression
暂无分享,去创建一个
Jian Huang | Yuling Jiao | Xiliang Lu | Yanyan Liu | Yuling Jiao | Xiliang Lu | Jian Huang | Yanyan Liu
[1] Lie Wang,et al. Orthogonal Matching Pursuit for Sparse Signal Recovery With Noise , 2011, IEEE Transactions on Information Theory.
[2] Tong Zhang. Some sharp performance bounds for least squares regression with L1 regularization , 2009, 0908.2869.
[3] Jianqing Fan,et al. Nonconcave penalized likelihood with a diverging number of parameters , 2004, math/0406466.
[4] Martin J. Wainwright,et al. Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$ -Constrained Quadratic Programming (Lasso) , 2009, IEEE Transactions on Information Theory.
[5] Yurii Nesterov,et al. Gradient methods for minimizing composite functions , 2012, Mathematical Programming.
[6] Stéphane Mallat,et al. Matching pursuits with time-frequency dictionaries , 1993, IEEE Trans. Signal Process..
[7] Tong Zhang,et al. Sparse Recovery With Orthogonal Matching Pursuit Under RIP , 2010, IEEE Transactions on Information Theory.
[8] Tong Zhang,et al. A General Theory of Concave Regularization for High-Dimensional Sparse Estimation Problems , 2011, 1108.4988.
[9] H. Zou,et al. STRONG ORACLE OPTIMALITY OF FOLDED CONCAVE PENALIZED ESTIMATION. , 2012, Annals of statistics.
[10] Zizhuo Wang,et al. Complexity of Unconstrained L2-Lp Minimization , 2011 .
[11] Yufeng Liu,et al. Variable Selection via A Combination of the L0 and L1 Penalties , 2007 .
[12] Xiaojun Chen,et al. Complexity of unconstrained $$L_2-L_p$$ minimization , 2011, Math. Program..
[13] T. Blumensath,et al. Iterative Thresholding for Sparse Approximations , 2008 .
[14] Martin J. Wainwright,et al. Minimax Rates of Estimation for High-Dimensional Linear Regression Over $\ell_q$ -Balls , 2009, IEEE Transactions on Information Theory.
[15] Y. She,et al. Thresholding-based iterative selection procedures for model selection and shrinkage , 2008, 0812.5061.
[16] Cun-Hui Zhang,et al. The sparsity and bias of the Lasso selection in high-dimensional linear regression , 2008, 0808.0967.
[17] Mike E. Davies,et al. Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.
[18] R. Tibshirani,et al. PATHWISE COORDINATE OPTIMIZATION , 2007, 0708.1485.
[19] Karim Lounici. Sup-norm convergence rate and sign concentration property of Lasso and Dantzig estimators , 2008, 0801.4610.
[20] Jin Liu,et al. A Unified Primal Dual Active Set Algorithm for Nonconvex Sparse Recovery , 2013 .
[21] Rahul Garg,et al. Gradient descent with sparsification: an iterative algorithm for sparse recovery with restricted isometry property , 2009, ICML '09.
[22] M. R. Osborne,et al. A new approach to variable selection in least squares problems , 2000 .
[23] Jianqing Fan,et al. Sure independence screening for ultrahigh dimensional feature space , 2006, math/0612857.
[24] Prateek Jain,et al. On Iterative Hard Thresholding Methods for High-dimensional M-Estimation , 2014, NIPS.
[25] Peng Zhao,et al. On Model Selection Consistency of Lasso , 2006, J. Mach. Learn. Res..
[26] Michael A. Saunders,et al. Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..
[27] Martin J. Wainwright,et al. Fast global convergence of gradient methods for high-dimensional statistical recovery , 2011, ArXiv.
[28] J. Horowitz,et al. Asymptotic properties of bridge estimators in sparse high-dimensional regression models , 2008, 0804.0693.
[29] N. Meinshausen,et al. High-dimensional graphs and variable selection with the Lasso , 2006, math/0608017.
[30] Yongdai Kim,et al. Smoothly Clipped Absolute Deviation on High Dimensions , 2008 .
[31] Stéphane Canu,et al. Recovering Sparse Signals With a Certain Family of Nonconvex Penalties and DC Programming , 2009, IEEE Transactions on Signal Processing.
[32] Yaakov Tsaig,et al. Fast Solution of $\ell _{1}$ -Norm Minimization Problems When the Solution May Be Sparse , 2008, IEEE Transactions on Information Theory.
[33] D. Hunter,et al. Optimization Transfer Using Surrogate Objective Functions , 2000 .
[34] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[35] R. Tibshirani,et al. Least angle regression , 2004, math/0406456.
[36] Cun-Hui Zhang,et al. Adaptive Lasso for sparse high-dimensional regression models , 2008 .
[37] Wenjiang J. Fu. Penalized Regressions: The Bridge versus the Lasso , 1998 .
[38] Po-Ling Loh,et al. Regularized M-estimators with nonconvexity: statistical and algorithmic theory for local optima , 2013, J. Mach. Learn. Res..
[39] Joel A. Tropp,et al. Greed is good: algorithmic results for sparse approximation , 2004, IEEE Transactions on Information Theory.
[40] T. Hastie,et al. SparseNet: Coordinate Descent With Nonconvex Penalties , 2011, Journal of the American Statistical Association.
[41] Jianqing Fan,et al. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .
[42] Jian Huang,et al. COORDINATE DESCENT ALGORITHMS FOR NONCONVEX PENALIZED REGRESSION, WITH APPLICATIONS TO BIOLOGICAL FEATURE SELECTION. , 2011, The annals of applied statistics.
[43] Tong Zhang,et al. Adaptive Forward-Backward Greedy Algorithm for Learning Sparse Representations , 2011, IEEE Transactions on Information Theory.
[44] K. Lange,et al. Coordinate descent algorithms for lasso penalized regression , 2008, 0803.3876.
[45] D. Bertsimas,et al. Best Subset Selection via a Modern Optimization Lens , 2015, 1507.03133.
[46] Zhaoran Wang,et al. OPTIMAL COMPUTATIONAL AND STATISTICAL RATES OF CONVERGENCE FOR SPARSE NONCONVEX LEARNING PROBLEMS. , 2013, Annals of statistics.
[47] Emmanuel J. Candès,et al. Decoding by linear programming , 2005, IEEE Transactions on Information Theory.
[48] E. Candès,et al. Stable signal recovery from incomplete and inaccurate measurements , 2005, math/0503066.
[49] Michael Elad,et al. Stable recovery of sparse overcomplete representations in the presence of noise , 2006, IEEE Transactions on Information Theory.
[50] Cun-Hui Zhang. Nearly unbiased variable selection under minimax concave penalty , 2010, 1002.4734.
[51] Tong Zhang,et al. Gradient Hard Thresholding Pursuit , 2018, J. Mach. Learn. Res..
[52] Bangti Jin,et al. A Primal Dual Active Set with Continuation Algorithm for the \ell^0-Regularized Optimization Problem , 2014, ArXiv.
[53] H. Zou. The Adaptive Lasso and Its Oracle Properties , 2006 .
[54] Balas K. Natarajan,et al. Sparse Approximate Solutions to Linear Systems , 1995, SIAM J. Comput..
[55] Lin Xiao,et al. A Proximal-Gradient Homotopy Method for the Sparse Least-Squares Problem , 2012, SIAM J. Optim..
[56] Tong Zhang,et al. Analysis of Multi-stage Convex Relaxation for Sparse Regularization , 2010, J. Mach. Learn. Res..
[57] Marc E. Pfetsch,et al. The Computational Complexity of the Restricted Isometry Property, the Nullspace Property, and Related Concepts in Compressed Sensing , 2012, IEEE Transactions on Information Theory.
[58] Runze Li,et al. CALIBRATING NON-CONVEX PENALIZED REGRESSION IN ULTRA-HIGH DIMENSION. , 2013, Annals of statistics.
[59] D. Hunter,et al. Variable Selection using MM Algorithms. , 2005, Annals of statistics.