Catalyst Acceleration for Gradient-Based Non-Convex Optimization
暂无分享,去创建一个
Z. Harchaoui | J. Mairal | D. Drusvyatskiy | Hongzhou Lin | Courtney Paquette | C. Paquette | Zaïd Harchaoui
[1] R. Rockafellar. Monotone Operators and the Proximal Point Algorithm , 1976 .
[2] R. Rockafellar. Favorable Classes of Lipschitz Continuous Functions in Subgradient Optimization , 1981 .
[3] Y. Nesterov. A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .
[4] Osman Güer. On the convergence of the proximal point algorithm for convex minimization , 1991 .
[5] Osman Güler,et al. New Proximal Point Algorithms for Convex Minimization , 1992, SIAM J. Optim..
[6] J. Hiriart-Urruty,et al. Convex analysis and minimization algorithms , 1993 .
[7] F. Clarke,et al. Proximal Smoothness and the Lower{C 2 Property , 1995 .
[8] R. Rockafellar,et al. Prox-regular functions in variational analysis , 1996 .
[9] J. Borwein,et al. Convex Analysis And Nonlinear Optimization , 2000 .
[10] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[11] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[12] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[13] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[14] Nicholas I. M. Gould,et al. On the Complexity of Steepest Descent, Newton's and Regularized Newton's Methods for Nonconvex Unconstrained Optimization Problems , 2010, SIAM J. Optim..
[15] Guillermo Sapiro,et al. Online Learning for Matrix Factorization and Sparse Coding , 2009, J. Mach. Learn. Res..
[16] Heinz H. Bauschke,et al. Convex Analysis and Monotone Operator Theory in Hilbert Spaces , 2011, CMS Books in Mathematics.
[17] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[18] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[19] Yurii Nesterov,et al. Gradient methods for minimizing composite functions , 2012, Mathematical Programming.
[20] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[21] Jean Ponce,et al. Sparse Modeling for Image and Vision Processing , 2014, Found. Trends Comput. Graph. Vis..
[22] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[23] Lin Xiao,et al. A Proximal Stochastic Gradient Method with Progressive Variance Reduction , 2014, SIAM J. Optim..
[24] Nicholas I. M. Gould,et al. On the complexity of finding first-order critical points in constrained nonlinear optimization , 2014, Math. Program..
[25] Bastian Goldlücke,et al. Variational Analysis , 2014, Computer Vision, A Reference Guide.
[26] Stephen P. Boyd,et al. Proximal Algorithms , 2013, Found. Trends Optim..
[27] Trevor Hastie,et al. Statistical Learning with Sparsity: The Lasso and Generalizations , 2015 .
[28] Huan Li,et al. Accelerated Proximal Gradient Methods for Nonconvex Programming , 2015, NIPS.
[29] Stephen J. Wright. Coordinate descent algorithms , 2015, Mathematical Programming.
[30] Mark W. Schmidt,et al. Coordinate Descent Converges Faster with the Gauss-Southwell Rule Than Random Selection , 2015, ICML.
[31] Un-regularizing: approximate proximal point algorithms for empirical risk minimization , 2015 .
[32] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[33] Julien Mairal,et al. Incremental Majorization-Minimization Optimization with Application to Large-Scale Machine Learning , 2014, SIAM J. Optim..
[34] Zaïd Harchaoui,et al. A Universal Catalyst for First-Order Optimization , 2015, NIPS.
[35] Dimitri P. Bertsekas,et al. Convex Optimization Algorithms , 2015 .
[36] Nathan Srebro,et al. Tight Complexity Bounds for Optimizing Composite Objectives , 2016, NIPS.
[37] Saeed Ghadimi,et al. Accelerated gradient methods for nonconvex nonlinear and stochastic programming , 2013, Mathematical Programming.
[38] Zeyuan Allen Zhu,et al. Improved SVRG for Non-Strongly-Convex or Sum-of-Non-Convex Objectives , 2015, ICML.
[39] Yair Carmon,et al. Accelerated Methods for Non-Convex Optimization , 2016, SIAM J. Optim..
[40] Alexander J. Smola,et al. Proximal Stochastic Methods for Nonsmooth Nonconvex Finite-Sum Optimization , 2016, NIPS.
[41] Zeyuan Allen Zhu,et al. Variance Reduction for Faster Non-Convex Optimization , 2016, ICML.
[42] Alexander J. Smola,et al. Stochastic Variance Reduction for Nonconvex Optimization , 2016, ICML.
[43] Shai Shalev-Shwartz,et al. SDCA without Duality, Regularization, and Individual Convexity , 2016, ICML.
[44] Zeyuan Allen-Zhu,et al. Katyusha: the first direct acceleration of stochastic gradient methods , 2016, J. Mach. Learn. Res..
[45] Zaïd Harchaoui,et al. Catalyst Acceleration for First-order Convex Optimization: from Theory to Practice , 2017, J. Mach. Learn. Res..
[46] Michael I. Jordan,et al. Non-convex Finite-Sum Optimization Via SCSG Methods , 2017, NIPS.
[47] Zeyuan Allen-Zhu,et al. Natasha: Faster Non-Convex Stochastic Optimization via Strongly Non-Convex Parameter , 2017, ICML.
[48] Yair Carmon,et al. "Convex Until Proven Guilty": Dimension-Free Acceleration of Gradient Descent on Non-Convex Functions , 2017, ICML.
[49] Bruce W. Suter,et al. From error bounds to the complexity of first-order descent methods for convex functions , 2015, Math. Program..
[50] Mark W. Schmidt,et al. Minimizing finite sums with the stochastic average gradient , 2013, Mathematical Programming.
[51] Michael O'Neill,et al. Behavior of Accelerated Gradient Methods Near Critical Points of Nonconvex Problems , 2017 .
[52] Zeyuan Allen-Zhu. Natasha: Faster Stochastic Non-Convex Optimization via Strongly Non-Convex Parameter , 2017 .
[53] Michael I. Jordan,et al. Less than a Single Pass: Stochastically Controlled Stochastic Gradient , 2016, AISTATS.
[54] Michael I. Jordan,et al. Accelerated Gradient Descent Escapes Saddle Points Faster than Gradient Descent , 2017, COLT.
[55] Yi Zhou,et al. An optimal randomized incremental gradient method , 2015, Mathematical Programming.
[56] Zeyuan Allen-Zhu,et al. How To Make the Gradients Small Stochastically: Even Faster Convex and Nonconvex SGD , 2018, NeurIPS.
[57] Yair Carmon,et al. Accelerated Methods for NonConvex Optimization , 2018, SIAM J. Optim..
[58] Stephen J. Wright,et al. Behavior of accelerated gradient methods near critical points of nonconvex functions , 2017, Math. Program..
[59] Saeed Ghadimi,et al. Generalized Uniformly Optimal Methods for Nonlinear Programming , 2015, Journal of Scientific Computing.
[60] Dmitriy Drusvyatskiy,et al. Efficiency of minimizing compositions of convex functions and smooth maps , 2016, Math. Program..
[61] K. Schittkowski,et al. NONLINEAR PROGRAMMING , 2022 .