Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition
暂无分享,去创建一个
[1] A. Hoffman. On approximate solutions of systems of linear inequalities , 1952 .
[2] N. Wormald. A 4-chromatic graph with a special plane drawing , 1979 .
[3] M. A. Hanson. On sufficiency of the Kuhn-Tucker conditions , 1981 .
[4] B. M. Glover,et al. Invex functions and duality , 1985, Journal of the Australian Mathematical Society. Series A. Pure Mathematics and Statistics.
[5] Adi Ben-Israel,et al. What is invexity? , 1986, The Journal of the Australian Mathematical Society. Series B. Applied Mathematics.
[6] Martin A. Riedmiller,et al. RPROP - A Fast Adaptive Learning Algorithm , 1992 .
[7] Paul Tseng,et al. Error Bound and Convergence Analysis of Matrix Splitting Algorithms for the Affine Variational Inequality Problem , 1992, SIAM J. Optim..
[8] Z.-Q. Luo,et al. Error bounds and convergence analysis of feasible descent methods: a general approach , 1993, Ann. Oper. Res..
[9] K. Kurdyka. On gradients of functions definable in o-minimal structures , 1998 .
[10] Mihai Anitescu,et al. Degenerate Nonlinear Programming with a Quadratic Growth Condition , 1999, SIAM J. Optim..
[11] Gunnar Rätsch,et al. An Introduction to Boosting and Leveraging , 2002, Machine Learning Summer School.
[12] Dimitri P. Bertsekas,et al. Convex Analysis and Optimization , 2003 .
[13] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[14] Don R. Hush,et al. QP Algorithms with Guaranteed Accuracy and Run Time for Support Vector Machines , 2006, J. Mach. Learn. Res..
[15] Yurii Nesterov,et al. Cubic regularization of Newton method and its global performance , 2006, Math. Program..
[16] J. Bolte,et al. Characterizations of Lojasiewicz inequalities and applications , 2008, 0802.0826.
[17] P. Tseng,et al. Block-Coordinate Gradient Descent Method for Linearly Constrained Nonsmooth Separable Optimization , 2009 .
[18] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[19] J. Bolte,et al. Characterizations of Lojasiewicz inequalities: Subgradient flows, talweg, convexity , 2009 .
[20] Hédy Attouch,et al. On the convergence of the proximal algorithm for nonsmooth functions involving analytic features , 2008, Math. Program..
[21] Alexander Shapiro,et al. Stochastic Approximation approach to Stochastic Programming , 2013 .
[22] Paul Tseng,et al. Approximation accuracy, gradient methods, and error bound for structured convex optimization , 2010, Math. Program..
[23] Martin J. Wainwright,et al. Fast global convergence rates of gradient methods for high-dimensional statistical recovery , 2010, NIPS.
[24] Martin J. Wainwright,et al. Fast global convergence of gradient methods for high-dimensional statistical recovery , 2011, ArXiv.
[25] Peter V. Gehler,et al. Learning Output Kernels with Block Coordinate Descent , 2011, ICML.
[26] Mark W. Schmidt,et al. Convergence Rates of Inexact Proximal-Gradient Methods for Convex Optimization , 2011, NIPS.
[27] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[28] Mark W. Schmidt,et al. A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets , 2012, NIPS.
[29] Ming Gu,et al. ParNes: a rapidly convergent algorithm for accurate recovery of sparse and approximately sparse signals , 2009, Numerical Algorithms.
[30] Zhi-Quan Luo,et al. On the Linear Convergence of the Proximal Gradient Method for Trace Norm Regularization , 2013, NIPS.
[31] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[32] Benar Fux Svaiter,et al. Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward–backward splitting, and regularized Gauss–Seidel methods , 2013, Math. Program..
[33] Lin Xiao,et al. A Proximal-Gradient Homotopy Method for the Sparse Least-Squares Problem , 2012, SIAM J. Optim..
[34] Eric Moulines,et al. Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n) , 2013, NIPS.
[35] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[36] Hui Zhang,et al. Gradient methods for convex minimization: better rates under weaker conditions , 2013, ArXiv.
[37] Z. Luo,et al. On the Linear Convergence of a Proximal Gradient Method for a Class of Nonsmooth Convex Minimization Problems , 2013 .
[38] Chih-Jen Lin,et al. Iteration complexity of feasible descent methods for convex optimization , 2014, J. Mach. Learn. Res..
[39] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[40] Pinghua Gong,et al. Linear Convergence of Variance-Reduced Stochastic Gradient without Strong Convexity , 2014 .
[41] Z. Luo,et al. On the Linear Convergence of the Approximate Proximal Splitting Method for Non-smooth Convex Optimization , 2014, 1404.5350.
[42] Stephen J. Wright,et al. An asynchronous parallel stochastic coordinate descent algorithm , 2013, J. Mach. Learn. Res..
[43] Mark W. Schmidt,et al. Coordinate Descent Converges Faster with the Gauss-Southwell Rule Than Random Selection , 2015, ICML.
[44] Elad Hazan,et al. Faster Rates for the Frank-Wolfe Method over Strongly-Convex Sets , 2014, ICML.
[45] Stephen J. Wright,et al. Asynchronous Stochastic Coordinate Descent: Parallelism and Convergence Properties , 2014, SIAM J. Optim..
[46] Elad Hazan,et al. Fast and Simple PCA via Convex Optimization , 2015, ArXiv.
[47] Alexander J. Smola,et al. Fast Stochastic Methods for Nonsmooth Nonconvex Optimization , 2016, ArXiv.
[48] Ion Necoara,et al. Parallel Random Coordinate Descent Method for Composite Minimization: Convergence Analysis and Error Bounds , 2016, SIAM J. Optim..
[49] Alexander J. Smola,et al. Fast Incremental Method for Nonconvex Optimization , 2016, ArXiv.
[50] Hui Zhang. Characterization of linear convergence of gradient descent , 2016 .
[51] Alexander J. Smola,et al. Fast incremental method for smooth nonconvex optimization , 2016, 2016 IEEE 55th Conference on Decision and Control (CDC).
[52] Alexander J. Smola,et al. Stochastic Frank-Wolfe methods for nonconvex optimization , 2016, 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[53] Alexander J. Smola,et al. Stochastic Variance Reduction for Nonconvex Optimization , 2016, ICML.
[54] Linear convergence of randomized feasible descent methods under the weak strong convexity assumption , 2016 .
[55] Hui Zhang,et al. The restricted strong convexity revisited: analysis of equivalence to error bound and quadratic growth , 2015, Optim. Lett..
[56] Bruce W. Suter,et al. From error bounds to the complexity of first-order descent methods for convex functions , 2015, Math. Program..
[57] Anthony Man-Cho So,et al. A unified approach to error bounds for structured convex optimization problems , 2015, Math. Program..
[58] Panagiotis Patrinos,et al. Forward–backward quasi-Newton methods for nonsmooth optimization problems , 2016, Comput. Optim. Appl..
[59] Guoyin Li,et al. Calculus of the Exponent of Kurdyka–Łojasiewicz Inequality and Its Applications to Linear Convergence of First-Order Methods , 2016, Foundations of Computational Mathematics.
[60] Dmitriy Drusvyatskiy,et al. Error Bounds, Quadratic Growth, and Linear Convergence of Proximal Methods , 2016, Math. Oper. Res..