Accelerated Doubly Stochastic Gradient Algorithm for Large-scale Empirical Risk Minimization
暂无分享,去创建一个
Chao Zhang | Zebang Shen | Hui Qian | Tongzhou Mu | Chao Zhang | Hui Qian | Zebang Shen | Tongzhou Mu
[1] L. Tippett,et al. Applied Statistics. A Journal of the Royal Statistical Society , 1952 .
[2] Mark W. Schmidt,et al. Minimizing finite sums with the stochastic average gradient , 2013, Mathematical Programming.
[3] Yiming Wang,et al. Accelerated Mini-batch Randomized Block Coordinate Descent Method , 2014, NIPS.
[4] Noah Simon,et al. A Sparse-Group Lasso , 2013 .
[5] Quanquan Gu,et al. Accelerated Stochastic Block Coordinate Descent with Optimal Sampling , 2016, KDD.
[6] Zeyuan Allen Zhu,et al. Katyusha: Accelerated Variance Reduction for Faster SGD , 2016, ArXiv.
[7] Atsushi Nitanda,et al. Stochastic Proximal Gradient Descent with Acceleration Techniques , 2014, NIPS.
[8] Peter Richtárik,et al. Accelerated, Parallel, and Proximal Coordinate Descent , 2013, SIAM J. Optim..
[9] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[10] Zaïd Harchaoui,et al. A Universal Catalyst for First-Order Optimization , 2015, NIPS.
[11] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[12] Yin Tat Lee,et al. Efficient Accelerated Coordinate Descent Methods and Faster Algorithms for Solving Linear Systems , 2013, 2013 IEEE 54th Annual Symposium on Foundations of Computer Science.
[13] James T. Kwok,et al. Accelerated Gradient Methods for Stochastic Optimization and Online Learning , 2009, NIPS.
[14] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[15] Stephen J. Wright. Coordinate descent algorithms , 2015, Mathematical Programming.
[16] Guanghui Lan,et al. An optimal method for stochastic composite optimization , 2012, Math. Program..
[17] Sham M. Kakade,et al. Un-regularizing: approximate proximal point and faster stochastic algorithms for empirical risk minimization , 2015, ICML.
[18] Nathan Srebro,et al. Tight Complexity Bounds for Optimizing Composite Objectives , 2016, NIPS.
[19] Y. Nesterov. A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .
[20] Tong Zhang,et al. Accelerated proximal stochastic dual coordinate ascent for regularized loss minimization , 2013, Math. Program..
[21] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[22] Yuchen Zhang,et al. Stochastic Primal-Dual Coordinate Method for Regularized Empirical Risk Minimization , 2014, ICML.
[23] Lin Xiao,et al. An Accelerated Randomized Proximal Coordinate Gradient Method and its Application to Regularized Empirical Risk Minimization , 2015, SIAM J. Optim..
[24] Trevor Hastie,et al. The Elements of Statistical Learning , 2001 .
[25] Boris Polyak. Some methods of speeding up the convergence of iteration methods , 1964 .
[26] Zeyuan Allen Zhu,et al. Linear Coupling: An Ultimate Unification of Gradient and Mirror Descent , 2014, ITCS.
[27] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[28] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[29] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.