暂无分享,去创建一个
Michael I. Jordan | Lihua Lei | Peter Richt'arik | Samuel Horv'ath | Peter Richtárik | Lihua Lei | Samuel Horváth
[1] Guanghui Lan,et al. A unified variance-reduced accelerated gradient method for convex optimization , 2019, NeurIPS.
[2] Boris Polyak. Some methods of speeding up the convergence of iteration methods , 1964 .
[3] Yi Zhou,et al. SpiderBoost: A Class of Faster Variance-reduced Algorithms for Nonconvex Optimization , 2018, ArXiv.
[4] Zeyuan Allen-Zhu,et al. How To Make the Gradients Small Stochastically: Even Faster Convex and Nonconvex SGD , 2018, NeurIPS.
[5] Peter Richtárik,et al. Semi-Stochastic Gradient Descent Methods , 2013, Front. Appl. Math. Stat..
[6] Aurélien Lucchi,et al. Variance Reduced Stochastic Gradient Descent with Neighbors , 2015, NIPS.
[7] Léon Bottou,et al. A Lower Bound for the Optimization of Finite Sums , 2014, ICML.
[8] Tianbao Yang,et al. Accelerate stochastic subgradient method by leveraging local growth condition , 2016, Analysis and Applications.
[9] Peter Richtárik,et al. One Method to Rule Them All: Variance Reduction for Data, Parameters and Many New Methods , 2019, ArXiv.
[10] Alexander J. Smola,et al. Stochastic Variance Reduction for Nonconvex Optimization , 2016, ICML.
[11] Quanquan Gu,et al. Stochastic Nested Variance Reduction for Nonconvex Optimization , 2018, J. Mach. Learn. Res..
[12] Jie Liu,et al. SARAH: A Novel Method for Machine Learning Problems Using Stochastic Recursive Gradient , 2017, ICML.
[13] F. Bach,et al. Stochastic quasi-gradient methods: variance reduction via Jacobian sketching , 2018, Mathematical Programming.
[14] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[15] Zeyuan Allen-Zhu,et al. Katyusha: the first direct acceleration of stochastic gradient methods , 2016, J. Mach. Learn. Res..
[16] Boris Polyak,et al. Acceleration of stochastic approximation by averaging , 1992 .
[17] S. Kakade,et al. Revisiting the Polyak step size , 2019, 1905.00313.
[18] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[19] Dan Alistarh,et al. QSGD: Communication-Optimal Stochastic Gradient Descent, with Applications to Training Neural Networks , 2016, 1610.02132.
[20] Tianbao Yang,et al. Adaptive SVRG Methods under Error Bound Conditions with Unknown Growth Parameter , 2017, NIPS.
[21] Dmitry Kovalev,et al. Stochastic Newton and Cubic Newton Methods with Simple Local Linear-Quadratic Rates , 2019, ArXiv.
[22] Michael I. Jordan,et al. Variance Reduction with Sparse Gradients , 2020, ICLR.
[23] Jian Li,et al. A Simple Proximal Stochastic Gradient Method for Nonsmooth Nonconvex Optimization , 2018, NeurIPS.
[24] Jakub Konecný,et al. S2CD: Semi-stochastic coordinate descent , 2014 .
[25] Yurii Nesterov,et al. Lectures on Convex Optimization , 2018 .
[26] Michael I. Jordan,et al. Non-convex Finite-Sum Optimization Via SCSG Methods , 2017, NIPS.
[27] Lam M. Nguyen,et al. Hybrid Stochastic Gradient Descent Algorithms for Stochastic Nonconvex Optimization , 2019, 1905.05920.
[28] Michael I. Jordan,et al. Stochastic Cubic Regularization for Fast Nonconvex Optimization , 2017, NeurIPS.
[29] Artin,et al. SARAH : A Novel Method for Machine Learning Problems Using Stochastic Recursive Gradient , 2017 .
[30] Peter Richtárik,et al. SGD: General Analysis and Improved Rates , 2019, ICML 2019.
[31] Konstantin Mishchenko,et al. Adaptive gradient descent without descent , 2019, ICML.
[32] Tong Zhang,et al. SPIDER: Near-Optimal Non-Convex Optimization via Stochastic Path Integrated Differential Estimator , 2018, NeurIPS.
[33] Enhong Chen,et al. SADAGRAD: Strongly Adaptive Stochastic Gradient Methods , 2018, ICML.
[34] Mark W. Schmidt,et al. A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets , 2012, NIPS.
[35] Michael I. Jordan,et al. Distributed optimization with arbitrary local solvers , 2015, Optim. Methods Softw..
[36] Michael I. Jordan,et al. On the Adaptivity of Stochastic Gradient-Based Optimization , 2019, SIAM J. Optim..
[37] Yurii Nesterov,et al. Universal gradient methods for convex optimization problems , 2015, Math. Program..
[38] Peter Richtárik,et al. Don't Jump Through Hoops and Remove Those Loops: SVRG and Katyusha are Better Without the Outer Loop , 2019, ALT.
[39] Sebastian U. Stich,et al. k-SVRG: Variance Reduction for Large Scale Optimization , 2018, 1805.00982.
[40] Justin Domke,et al. Finito: A faster, permutable incremental gradient method for big data problems , 2014, ICML.
[41] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[42] Yi Zhou,et al. SpiderBoost and Momentum: Faster Stochastic Variance Reduction Algorithms , 2018 .
[43] Marten van Dijk,et al. Finite-sum smooth optimization with SARAH , 2019, Computational Optimization and Applications.
[44] Ohad Shamir,et al. Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.
[45] John Darzentas,et al. Problem Complexity and Method Efficiency in Optimization , 1983 .
[46] Sebastian U. Stich,et al. Local SGD Converges Fast and Communicates Little , 2018, ICLR.
[47] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[48] Sanjiv Kumar,et al. On the Convergence of Adam and Beyond , 2018 .
[49] Michael I. Jordan,et al. Less than a Single Pass: Stochastically Controlled Stochastic Gradient , 2016, AISTATS.
[50] Peter Richtárik,et al. A Unified Theory of SGD: Variance Reduction, Sampling, Quantization and Coordinate Descent , 2019, AISTATS.
[51] Aaron Mishkin,et al. Painless Stochastic Gradient: Interpolation, Line-Search, and Convergence Rates , 2019, NeurIPS.
[52] Alexander Shapiro,et al. Stochastic Approximation approach to Stochastic Programming , 2013 .
[53] Eric Moulines,et al. Non-Asymptotic Analysis of Stochastic Approximation Algorithms for Machine Learning , 2011, NIPS.
[54] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[55] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[56] Volkan Cevher,et al. Online Adaptive Methods, Universality and Acceleration , 2018, NeurIPS.
[57] Eric Moulines,et al. Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n) , 2013, NIPS.
[58] Peter Richtárik,et al. Quartz: Randomized Dual Coordinate Ascent with Arbitrary Sampling , 2015, NIPS.
[59] Tong Zhang,et al. Stochastic Optimization with Importance Sampling for Regularized Loss Minimization , 2014, ICML.
[60] D. Ruppert,et al. Efficient Estimations from a Slowly Convergent Robbins-Monro Process , 1988 .
[61] Francis R. Bach,et al. From Averaging to Acceleration, There is Only a Step-size , 2015, COLT.