Adaptive SVRG Methods under Error Bound Conditions with Unknown Growth Parameter
暂无分享,去创建一个
Tianbao Yang | Yi Xu | Qihang Lin | Tianbao Yang | Yi Xu | Qihang Lin
[1] R. Tyrrell Rockafellar,et al. Convex Analysis , 1970, Princeton Landmarks in Mathematics and Physics.
[2] H. Nyquist. The optimal Lp norm estimator in linear regression models , 1983 .
[3] P. Tseng,et al. On the linear convergence of descent methods for convex essentially smooth minimization , 1992 .
[4] P. Tseng,et al. On the convergence of the coordinate descent method for convex differentiable minimization , 1992 .
[5] Z.-Q. Luo,et al. Error bounds and convergence analysis of feasible descent methods: a general approach , 1993, Ann. Oper. Res..
[6] Zhi-Quan Luo,et al. On the Linear Convergence of the Proximal Gradient Method for Trace Norm Regularization , 2013, NIPS.
[7] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[8] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[9] Guoyin Li,et al. Global error bounds for piecewise convex polynomials , 2013, Math. Program..
[10] Hui Zhang,et al. Gradient methods for convex minimization: better rates under weaker conditions , 2013, ArXiv.
[11] Yurii Nesterov,et al. Gradient methods for minimizing composite functions , 2012, Mathematical Programming.
[12] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[13] Lin Xiao,et al. A Proximal Stochastic Gradient Method with Progressive Variance Reduction , 2014, SIAM J. Optim..
[14] Jieping Ye,et al. Linear Convergence of Variance-Reduced Projected Stochastic Gradient without Strong Convexity , 2014, ArXiv.
[15] Tianbao Yang,et al. RSG: Beating SGD without Smoothness and/or Strong Convexity , 2015 .
[16] Lin Xiao,et al. An adaptive accelerated proximal gradient method and its homotopy continuation for sparse optimization , 2014, Computational Optimization and Applications.
[17] Stephen J. Wright,et al. Asynchronous Stochastic Coordinate Descent: Parallelism and Convergence Properties , 2014, SIAM J. Optim..
[18] Qi Zhang,et al. \(\ell_{1, p}\)-Norm Regularization: Error Bounds and Convergence Rate Analysis of First-Order Methods , 2015, ICML.
[19] Mark W. Schmidt,et al. Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition , 2016, ECML/PKDD.
[20] Tianbao Yang,et al. Homotopy Smoothing for Non-Smooth Problems with Lower Complexity than O(1/\epsilon) , 2016, NIPS.
[21] Zeyuan Allen Zhu,et al. Improved SVRG for Non-Strongly-Convex or Sum-of-Non-Convex Objectives , 2015, ICML.
[22] Martin Takác,et al. Linear Convergence of Randomized Feasible Descent Methods Under the Weak Strong Convexity Assumption , 2015, J. Mach. Learn. Res..
[23] Jie Liu,et al. Projected Semi-Stochastic Gradient Descent Method with Mini-Batch Scheme under Weak Strong Convexity Assumption , 2016, ArXiv.
[24] Mingrui Liu,et al. Adaptive Accelerated Gradient Converging Method under H\"{o}lderian Error Bound Condition , 2016, NIPS.
[25] Taiji Suzuki,et al. Doubly Accelerated Stochastic Variance Reduced Dual Averaging Method for Regularized Empirical Risk Minimization , 2017, NIPS.
[26] Bruce W. Suter,et al. From error bounds to the complexity of first-order descent methods for convex functions , 2015, Math. Program..
[27] Anthony Man-Cho So,et al. A unified approach to error bounds for structured convex optimization problems , 2015, Mathematical Programming.
[28] Jie Liu,et al. SARAH: A Novel Method for Machine Learning Problems Using Stochastic Recursive Gradient , 2017, ICML.
[29] Tianbao Yang,et al. Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence , 2017, ICML.
[30] Zeyuan Allen Zhu,et al. Katyusha: the first direct acceleration of stochastic gradient methods , 2017, STOC.
[31] Dmitriy Drusvyatskiy,et al. Error Bounds, Quadratic Growth, and Linear Convergence of Proximal Methods , 2016, Math. Oper. Res..
[32] Yurii Nesterov,et al. Linear convergence of first order methods for non-strongly convex optimization , 2015, Math. Program..