An asynchronous mini-batch algorithm for regularized stochastic optimization
暂无分享,去创建一个
Hamid Reza Feyzmahdavian | Mikael Johansson | Arda Aytekin | H. Feyzmahdavian | M. Johansson | Arda Aytekin
[1] John C. Duchi,et al. Distributed delayed stochastic optimization , 2011, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[2] Matthew J. Streeter,et al. Delay-Tolerant Algorithms for Asynchronous Distributed Online Learning , 2014, NIPS.
[3] Alexander J. Smola,et al. Parallelized Stochastic Gradient Descent , 2010, NIPS.
[4] Christopher Ré,et al. Parallel stochastic gradient algorithms for large-scale matrix completion , 2013, Mathematical Programming Computation.
[5] Aryan Mokhtari,et al. An approximate Newton method for distributed optimization , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[6] Martin J. Wainwright,et al. Randomized Smoothing for Stochastic Optimization , 2011, SIAM J. Optim..
[7] John N. Tsitsiklis,et al. Distributed Asynchronous Deterministic and Stochastic Gradient Optimization Algorithms , 1984, 1984 American Control Conference.
[8] John Langford,et al. Slow Learners are Fast , 2009, NIPS.
[9] Hamid Reza Feyzmahdavian,et al. An asynchronous mini-batch algorithm for regularized stochastic optimization , 2015, CDC.
[10] Alexander Shapiro,et al. Stochastic Approximation approach to Stochastic Programming , 2013 .
[11] Thomas Hofmann,et al. Communication-Efficient Distributed Dual Coordinate Ascent , 2014, NIPS.
[12] Aaron Q. Li,et al. Parameter Server for Distributed Machine Learning , 2013 .
[13] Yijun Huang,et al. Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization , 2015, NIPS.
[14] H. Robbins. A Stochastic Approximation Method , 1951 .
[15] Ming Yan,et al. ARock: an Algorithmic Framework for Asynchronous Parallel Coordinate Updates , 2015, SIAM J. Sci. Comput..
[16] Damiano Varagnolo,et al. Distributed quadratic programming under asynchronous and lossy communications via Newton-Raphson consensus , 2015, 2015 European Control Conference (ECC).
[17] Yuchen Zhang,et al. Splash: User-friendly Programming Interface for Parallelizing Stochastic Algorithms , 2015, ArXiv.
[18] James T. Kwok,et al. Accelerated Gradient Methods for Stochastic Optimization and Online Learning , 2009, NIPS.
[19] Paul Tseng,et al. Approximation accuracy, gradient methods, and error bound for structured convex optimization , 2010, Math. Program..
[20] Marc Teboulle,et al. Mirror descent and nonlinear projected subgradient methods for convex optimization , 2003, Oper. Res. Lett..
[21] Stephen J. Wright,et al. An asynchronous parallel stochastic coordinate descent algorithm , 2013, J. Mach. Learn. Res..
[22] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[23] Stephen J. Wright,et al. Asynchronous Stochastic Coordinate Descent: Parallelism and Convergence Properties , 2014, SIAM J. Optim..
[24] R. Rockafellar,et al. On the interchange of subdifferentiation and conditional expectation for convex functionals , 1982 .
[25] Heinz H. Bauschke,et al. Convex Analysis and Monotone Operator Theory in Hilbert Spaces , 2011, CMS Books in Mathematics.
[26] Michael G. Rabbat,et al. Distributed dual averaging for convex optimization under communication delays , 2012, 2012 American Control Conference (ACC).
[27] Pascal Bianchi,et al. Convergence of a Multi-Agent Projected Stochastic Gradient Algorithm for Non-Convex Optimization , 2011, IEEE Transactions on Automatic Control.
[28] Peter Richtárik,et al. Parallel coordinate descent methods for big data optimization , 2012, Mathematical Programming.
[29] Trishul M. Chilimbi,et al. Project Adam: Building an Efficient and Scalable Deep Learning Training System , 2014, OSDI.
[30] Hamid Reza Feyzmahdavian,et al. Asynchronous incremental block-coordinate descent , 2014, 2014 52nd Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[31] Mehran Mesbahi,et al. Online Distributed ADMM on Networks , 2014, ArXiv.
[32] Angelia Nedic,et al. On Stochastic Subgradient Mirror-Descent Algorithm with Weighted Averaging , 2013, SIAM J. Optim..
[33] O. Mangasarian,et al. Robust linear programming discrimination of two linearly inseparable sets , 1992 .
[34] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[35] Yurii Nesterov,et al. Primal-dual subgradient methods for convex problems , 2005, Math. Program..
[36] Xi Chen,et al. Optimal Regularized Dual Averaging Methods for Stochastic Optimization , 2012, NIPS.
[37] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[38] Deanna Needell,et al. Stochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithm , 2013, Mathematical Programming.
[39] Ambuj Tewari,et al. Stochastic methods for l1 regularized loss minimization , 2009, ICML '09.
[40] Eric R. Ziegel,et al. The Elements of Statistical Learning , 2003, Technometrics.
[41] Alexander J. Smola,et al. On Variance Reduction in Stochastic Gradient Descent and its Asynchronous Variants , 2015, NIPS.
[42] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[43] Hamid Reza Feyzmahdavian,et al. A delayed proximal gradient method with linear convergence rate , 2014, 2014 IEEE International Workshop on Machine Learning for Signal Processing (MLSP).
[44] John C. Duchi,et al. Asynchronous stochastic convex optimization , 2015, 1508.00882.
[45] James T. Kwok,et al. Asynchronous Distributed ADMM for Consensus Optimization , 2014, ICML.
[46] Zi Wang,et al. An Asynchronous Distributed Proximal Gradient Method for Composite Convex Optimization , 2014, ICML.
[47] Ambuj Tewari,et al. Composite objective mirror descent , 2010, COLT 2010.
[48] Stephen J. Wright. Coordinate descent algorithms , 2015, Mathematical Programming.
[49] Guanghui Lan,et al. An optimal method for stochastic composite optimization , 2011, Mathematical Programming.
[50] Saeed Ghadimi,et al. Optimal Stochastic Approximation Algorithms for Strongly Convex Stochastic Composite Optimization, II: Shrinking Procedures and Optimal Algorithms , 2013, SIAM J. Optim..
[51] Saeed Ghadimi,et al. Optimal Stochastic Approximation Algorithms for Strongly Convex Stochastic Composite Optimization I: A Generic Algorithmic Framework , 2012, SIAM J. Optim..
[52] Ohad Shamir,et al. Better Mini-Batch Algorithms via Accelerated Gradient Methods , 2011, NIPS.
[53] Kunle Olukotun,et al. Taming the Wild: A Unified Analysis of Hogwild-Style Algorithms , 2015, NIPS.
[54] John N. Tsitsiklis,et al. Parallel and distributed computation , 1989 .
[55] Yiming Yang,et al. RCV1: A New Benchmark Collection for Text Categorization Research , 2004, J. Mach. Learn. Res..
[56] Michael I. Jordan,et al. Ergodic mirror descent , 2011, 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[57] Vivek S. Borkar,et al. Distributed Asynchronous Incremental Subgradient Methods , 2001 .
[58] Peter Richtárik,et al. Distributed Block Coordinate Descent for Minimizing Partially Separable Functions , 2014, 1406.0238.
[59] Marc Teboulle,et al. Convergence Analysis of a Proximal-Like Minimization Algorithm Using Bregman Functions , 1993, SIAM J. Optim..
[60] Michael G. Rabbat,et al. Communication/Computation Tradeoffs in Consensus-Based Distributed Optimization , 2012, NIPS.
[61] Ohad Shamir,et al. Optimal Distributed Online Prediction Using Mini-Batches , 2010, J. Mach. Learn. Res..
[62] Ohad Shamir,et al. Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.
[63] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[64] Yuchen Zhang,et al. Communication-Efficient Distributed Optimization of Self-Concordant Empirical Loss , 2015, ArXiv.
[65] Inderjit S. Dhillon,et al. PASSCoDe: Parallel ASynchronous Stochastic dual Co-ordinate Descent , 2015, ICML.
[66] Lin Xiao,et al. Dual Averaging Methods for Regularized Stochastic Learning and Online Optimization , 2009, J. Mach. Learn. Res..
[67] Asuman E. Ozdaglar,et al. Distributed Subgradient Methods for Convex Optimization Over Random Networks , 2011, IEEE Transactions on Automatic Control.