On Quadratic Convergence of DC Proximal Newton Algorithm in Nonconvex Sparse Learning
暂无分享,去创建一个
Tong Zhang | Lin Yang | Lin F. Yang | Tuo Zhao | Jarvis D. Haupt | Xingguo Li | Jason Ge | T. Zhao | Xingguo Li | J. Haupt | T. Zhang | J. Ge
[1] Larry A. Wasserman,et al. The huge Package for High-dimensional Undirected Graph Estimation in R , 2012, J. Mach. Learn. Res..
[2] Lin Xiao,et al. A Proximal-Gradient Homotopy Method for the Sparse Least-Squares Problem , 2012, SIAM J. Optim..
[3] Trevor Hastie,et al. Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.
[4] Tong Zhang,et al. Analysis of Multi-stage Convex Relaxation for Sparse Regularization , 2010, J. Mach. Learn. Res..
[5] Jianqing Fan,et al. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .
[6] A. Belloni,et al. Square-Root Lasso: Pivotal Recovery of Sparse Signals via Conic Programming , 2011 .
[7] Steve R. Gunn,et al. Result Analysis of the NIPS 2003 Feature Selection Challenge , 2004, NIPS.
[8] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[9] Jean Ponce,et al. Sparse Modeling for Image and Vision Processing , 2014, Found. Trends Comput. Graph. Vis..
[10] Tong Zhang. Multi-stage Convex Relaxation for Feature Selection , 2011, 1106.0565.
[11] Tianqi Zhao,et al. A Likelihood Ratio Framework for High Dimensional Semiparametric Regression , 2014 .
[12] Marc Teboulle,et al. Fast Gradient-Based Algorithms for Constrained Total Variation Image Denoising and Deblurring Problems , 2009, IEEE Transactions on Image Processing.
[13] Pradeep Ravikumar,et al. Constant Nullspace Strong Convexity and Fast Convergence of Proximal Methods under High-Dimensional Settings , 2014, NIPS.
[14] R. Tibshirani,et al. Strong rules for discarding predictors in lasso‐type problems , 2010, Journal of the Royal Statistical Society. Series B, Statistical methodology.
[15] Tuo Zhao,et al. A First Order Free Lunch for SQRT-Lasso , 2016, ArXiv.
[16] S. Geer. HIGH-DIMENSIONAL GENERALIZED LINEAR MODELS AND THE LASSO , 2008, 0804.0703.
[17] S. Shalev-Shwartz,et al. Stochastic methods for {\it l}$_{\mbox{1}}$ regularized loss minimization , 2009, ICML 2009.
[18] Yurii Nesterov,et al. Gradient methods for minimizing composite functions , 2012, Mathematical Programming.
[19] M. B. Nebel,et al. Automated diagnoses of attention deficit hyperactive disorder using magnetic resonance imaging , 2012, Front. Syst. Neurosci..
[20] Shuheng Zhou. Restricted Eigenvalue Conditions on Subgaussian Random Matrices , 2009, 0912.4045.
[21] Po-Ling Loh,et al. Regularized M-estimators with nonconvexity: statistical and algorithmic theory for local optima , 2013, J. Mach. Learn. Res..
[22] Martin J. Wainwright,et al. A unified framework for high-dimensional analysis of $M$-estimators with decomposable regularizers , 2009, NIPS.
[23] Nicol N. Schraudolph,et al. A Fast, Compact Approximation of the Exponential Function , 1999, Neural Computation.
[24] Tuo Zhao,et al. Stochastic Variance Reduced Optimization for Nonconvex Sparse Learning , 2016, ICML.
[25] L. Armijo. Minimization of functions having Lipschitz continuous first partial derivatives. , 1966 .
[26] Martin J. Wainwright,et al. Restricted Eigenvalue Properties for Correlated Gaussian Designs , 2010, J. Mach. Learn. Res..
[27] Tuo Zhao,et al. Pathwise Coordinate Optimization for Sparse Learning: Algorithm and Theory , 2014, ArXiv.
[28] P. McCullagh,et al. Generalized Linear Models , 1984 .
[29] P. Tseng,et al. On the linear convergence of descent methods for convex essentially smooth minimization , 1992 .
[30] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[31] H. Zou,et al. An Efficient Algorithm for Computing the HHSVM and Its Generalizations , 2013 .
[32] Roummel F. Marcia,et al. Compressed Sensing Performance Bounds Under Poisson Noise , 2009, IEEE Transactions on Signal Processing.
[33] Jianqing Fan,et al. I-LAMM FOR SPARSE LEARNING: SIMULTANEOUS CONTROL OF ALGORITHMIC COMPLEXITY AND STATISTICAL ERROR. , 2015, Annals of statistics.
[34] Cun-Hui Zhang. Nearly unbiased variable selection under minimax concave penalty , 2010, 1002.4734.
[35] P. Bickel,et al. SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR , 2008, 0801.1095.
[36] Evan T. Geller,et al. Patterns and rates of exonic de novo mutations in autism spectrum disorders , 2012, Nature.
[37] S. Geer,et al. On the conditions used to prove oracle results for the Lasso , 2009, 0910.0722.
[38] Zhaoran Wang,et al. OPTIMAL COMPUTATIONAL AND STATISTICAL RATES OF CONVERGENCE FOR SPARSE NONCONVEX LEARNING PROBLEMS. , 2013, Annals of statistics.
[39] 秀俊 松井,et al. Statistics for High-Dimensional Data: Methods, Theory and Applications , 2014 .
[40] J. Pfanzagl. Parametric Statistical Theory , 1994 .
[41] Tuo Zhao,et al. The picasso Package for Nonconvex Regularized M-estimation in High Dimensions in R , 2015 .
[42] Michael A. Saunders,et al. Proximal Newton-Type Methods for Minimizing Composite Functions , 2012, SIAM J. Optim..