暂无分享,去创建一个
Alexander J. Smola | Barnabás Póczos | Suvrit Sra | Sashank J. Reddi | Alex Smola | B. Póczos | S. Sra
[1] Boris Polyak. Gradient methods for the minimisation of functionals , 1963 .
[2] Harold J. Kushner,et al. wchastic. approximation methods for constrained and unconstrained systems , 1978 .
[3] John N. Tsitsiklis,et al. Parallel and distributed computation , 1989 .
[4] John N. Tsitsiklis,et al. Parallel and distributed computation , 1989 .
[5] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[6] Yurii Nesterov,et al. Cubic regularization of Newton method and its global performance , 2006, Math. Program..
[7] H. Robbins. A Stochastic Approximation Method , 1951 .
[8] Alexander Shapiro,et al. Stochastic Approximation approach to Stochastic Programming , 2013 .
[9] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[10] I. Gijbels,et al. Penalized likelihood regression for generalized linear models with non-quadratic penalties , 2011 .
[11] John C. Duchi,et al. Distributed delayed stochastic optimization , 2011, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[12] Suvrit Sra,et al. Scalable nonconvex inexact proximal splitting , 2012, NIPS.
[13] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[14] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[15] Saeed Ghadimi,et al. Stochastic First- and Zeroth-Order Methods for Nonconvex Stochastic Programming , 2013, SIAM J. Optim..
[16] Daniela M. Witten,et al. An Introduction to Statistical Learning: with Applications in R , 2013 .
[17] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[18] Dong Yu,et al. Deep Learning: Methods and Applications , 2014, Found. Trends Signal Process..
[19] Ohad Shamir,et al. A Stochastic PCA and SVD Algorithm with an Exponential Convergence Rate , 2014, ICML.
[20] Léon Bottou,et al. A Lower Bound for the Optimization of Finite Sums , 2014, ICML.
[21] Shai Shalev-Shwartz,et al. SDCA without Duality , 2015, ArXiv.
[22] Dimitri P. Bertsekas,et al. Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey , 2015, ArXiv.
[23] Alexander J. Smola,et al. On Variance Reduction in Stochastic Gradient Descent and its Asynchronous Variants , 2015, NIPS.
[24] Prateek Jain,et al. Accelerating Stochastic Gradient Descent , 2017, ArXiv.
[25] Mark W. Schmidt,et al. Minimizing finite sums with the stochastic average gradient , 2013, Mathematical Programming.
[26] Mingyi Hong,et al. A Distributed, Asynchronous, and Incremental Algorithm for Nonconvex Optimization: An ADMM Approach , 2014, IEEE Transactions on Control of Network Systems.