Ju n 20 18 Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Lojasiewicz Condition
暂无分享,去创建一个
[1] Martin A. Riedmiller,et al. RPROP - A Fast Adaptive Learning Algorithm , 1992 .
[2] Z.-Q. Luo,et al. Error bounds and convergence analysis of feasible descent methods: a general approach , 1993, Ann. Oper. Res..
[3] Mihai Anitescu,et al. Degenerate Nonlinear Programming with a Quadratic Growth Condition , 1999, SIAM J. Optim..
[4] Dimitri P. Bertsekas,et al. Convex Analysis and Optimization , 2003 .
[5] Yurii Nesterov,et al. Cubic regularization of Newton method and its global performance , 2006, Math. Program..
[6] P. Tseng,et al. Block-Coordinate Gradient Descent Method for Linearly Constrained Nonsmooth Separable Optimization , 2009 .
[7] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[8] J. Bolte,et al. Characterizations of Lojasiewicz inequalities: Subgradient flows, talweg, convexity , 2009 .
[9] Hédy Attouch,et al. On the convergence of the proximal algorithm for nonsmooth functions involving analytic features , 2008, Math. Program..
[10] Paul Tseng,et al. Approximation accuracy, gradient methods, and error bound for structured convex optimization , 2010, Math. Program..
[11] Martin J. Wainwright,et al. Fast global convergence rates of gradient methods for high-dimensional statistical recovery , 2010, NIPS.
[12] Peter V. Gehler,et al. Learning Output Kernels with Block Coordinate Descent , 2011, ICML.
[13] Mark W. Schmidt,et al. Convergence Rates of Inexact Proximal-Gradient Methods for Convex Optimization , 2011, NIPS.
[14] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[15] Zhi-Quan Luo,et al. On the Linear Convergence of the Proximal Gradient Method for Trace Norm Regularization , 2013, NIPS.
[16] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[17] Benar Fux Svaiter,et al. Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward–backward splitting, and regularized Gauss–Seidel methods , 2013, Math. Program..
[18] Lin Xiao,et al. A Proximal-Gradient Homotopy Method for the Sparse Least-Squares Problem , 2012, SIAM J. Optim..
[19] Eric Moulines,et al. Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n) , 2013, NIPS.
[20] Hui Zhang,et al. Gradient methods for convex minimization: better rates under weaker conditions , 2013, ArXiv.
[21] Z. Luo,et al. On the Linear Convergence of a Proximal Gradient Method for a Class of Nonsmooth Convex Minimization Problems , 2013, Journal of the Operations Research Society of China.
[22] Chih-Jen Lin,et al. Iteration complexity of feasible descent methods for convex optimization , 2014, J. Mach. Learn. Res..
[23] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[24] Mark W. Schmidt,et al. Coordinate Descent Converges Faster with the Gauss-Southwell Rule Than Random Selection , 2015, ICML.
[25] Stephen J. Wright,et al. Asynchronous Stochastic Coordinate Descent: Parallelism and Convergence Properties , 2014, SIAM J. Optim..
[26] Elad Hazan,et al. Fast and Simple PCA via Convex Optimization , 2015, ArXiv.
[27] Alexander J. Smola,et al. Fast Stochastic Methods for Nonsmooth Nonconvex Optimization , 2016, ArXiv.
[28] Alexander J. Smola,et al. Fast Incremental Method for Nonconvex Optimization , 2016, ArXiv.
[29] Martin Takác,et al. Linear Convergence of Randomized Feasible Descent Methods Under the Weak Strong Convexity Assumption , 2015, J. Mach. Learn. Res..
[30] Alexander J. Smola,et al. Stochastic Variance Reduction for Nonconvex Optimization , 2016, ICML.
[31] Hui Zhang,et al. The restricted strong convexity revisited: analysis of equivalence to error bound and quadratic growth , 2015, Optim. Lett..
[32] Anthony Man-Cho So,et al. A unified approach to error bounds for structured convex optimization problems , 2015, Mathematical Programming.
[33] Panagiotis Patrinos,et al. Forward–backward quasi-Newton methods for nonsmooth optimization problems , 2016, Computational Optimization and Applications.
[34] Hui Zhang. New analysis of linear convergence of gradient-type methods via unifying error bound conditions , 2020, Math. Program..