暂无分享,去创建一个
[1] Yurii Nesterov,et al. Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems , 2012, SIAM J. Optim..
[2] Michael H. Bowling,et al. A Randomized Mirror Descent Algorithm for Large Scale Multiple Kernel Learning , 2012, ICML.
[3] Marc Teboulle,et al. Mirror descent and nonlinear projected subgradient methods for convex optimization , 2003, Oper. Res. Lett..
[4] Peter Richtárik,et al. Coordinate descent with arbitrary sampling II: expected separable overapproximation , 2014, Optim. Methods Softw..
[5] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[6] Francis R. Bach,et al. Stochastic Composite Least-Squares Regression with Convergence Rate $O(1/n)$ , 2017, COLT.
[7] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[8] Saeed Ghadimi,et al. Optimal Stochastic Approximation Algorithms for Strongly Convex Stochastic Composite Optimization I: A Generic Algorithmic Framework , 2012, SIAM J. Optim..
[9] Shai Ben-David,et al. Understanding Machine Learning: From Theory to Algorithms , 2014 .
[10] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[11] Haihao Lu. “Relative Continuity” for Non-Lipschitz Nonsmooth Convex Optimization Using Stochastic (or Deterministic) Mirror Descent , 2017, INFORMS Journal on Optimization.
[12] I. Csiszár. Why least squares and maximum entropy? An axiomatic approach to inference for linear inverse problems , 1991 .
[13] H. Robbins. A Stochastic Approximation Method , 1951 .
[14] Alexandre M. Bayen,et al. Accelerated Mirror Descent in Continuous and Discrete Time , 2015, NIPS.
[15] Peter Richtárik,et al. Coordinate descent with arbitrary sampling I: algorithms and complexity† , 2014, Optim. Methods Softw..
[16] John Darzentas,et al. Problem Complexity and Method Efficiency in Optimization , 1983 .
[17] Li Zhang,et al. Proportional response dynamics in the Fisher market , 2009, Theor. Comput. Sci..
[18] Kenneth Lange,et al. MM optimization algorithms , 2016 .
[19] Ohad Shamir,et al. Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.
[20] Martin Benning,et al. Gradient descent in a generalised Bregman distance framework , 2016 .
[21] Marc Teboulle,et al. A Descent Lemma Beyond Lipschitz Gradient Continuity: First-Order Methods Revisited and Applications , 2017, Math. Oper. Res..
[22] O. Nelles,et al. An Introduction to Optimization , 1996, IEEE Antennas and Propagation Magazine.
[23] Yurii Nesterov,et al. Relatively Smooth Convex Optimization by First-Order Methods, and Applications , 2016, SIAM J. Optim..
[24] Y. Nesterov. A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .
[25] Cuong V Nguyen,et al. Accelerated Stochastic Mirror Descent Algorithms For Composite Non-strongly Convex Optimization , 2016 .
[26] Angelia Nedic,et al. On Stochastic Subgradient Mirror-Descent Algorithm with Weighted Averaging , 2013, SIAM J. Optim..
[27] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[28] Alexander Shapiro,et al. Stochastic Approximation approach to Stochastic Programming , 2013 .
[29] Guanghui Lan,et al. Primal-dual first-order methods with O (1/e) iteration-complexity for cone programming. , 2011 .
[30] Zeyuan Allen Zhu,et al. Linear Coupling: An Ultimate Unification of Gradient and Mirror Descent , 2014, ITCS.
[31] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[32] Peter Richtárik,et al. Parallel coordinate descent methods for big data optimization , 2012, Mathematical Programming.
[33] Peter Richtárik,et al. On optimal probabilities in stochastic coordinate descent methods , 2013, Optim. Lett..
[34] M. Bertero,et al. Image deblurring with Poisson data: from cells to galaxies , 2009 .
[35] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[36] Jie Liu,et al. SARAH: A Novel Method for Machine Learning Problems Using Stochastic Recursive Gradient , 2017, ICML.
[37] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[38] Peter Richtárik,et al. On the complexity of parallel coordinate descent , 2015, Optim. Methods Softw..
[39] Mark W. Schmidt,et al. A Stochastic Gradient Method with an Exponential Convergence Rate for Finite Training Sets , 2012, NIPS.
[40] Guanghui Lan,et al. Stochastic Block Mirror Descent Methods for Nonsmooth and Stochastic Optimization , 2013, SIAM J. Optim..
[41] Nikhil R. Devanur,et al. Distributed algorithms via gradient descent for fisher markets , 2011, EC '11.