Communication-Efficient Distributed Dual Coordinate Ascent
暂无分享,去创建一个
Thomas Hofmann | Michael I. Jordan | Martin Jaggi | Martin Takác | Sanjay Krishnan | Virginia Smith | Jonathan Terhorst | Martin Jaggi | S. Krishnan | Thomas Hofmann | Jonathan Terhorst | Virginia Smith | Martin Takác
[1] Stephen J. Wright,et al. An asynchronous parallel stochastic coordinate descent algorithm , 2013, J. Mach. Learn. Res..
[2] John C. Duchi,et al. Distributed delayed stochastic optimization , 2011, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[3] Peter Richtárik,et al. Fast distributed coordinate descent for non-strongly convex losses , 2014, 2014 IEEE International Workshop on Machine Learning for Signal Processing (MLSP).
[4] Peter Richtárik,et al. Distributed Coordinate Descent Method for Learning with Big Data , 2013, J. Mach. Learn. Res..
[5] Martin J. Wainwright,et al. Communication-efficient algorithms for statistical optimization , 2012, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[6] Tianbao Yang,et al. Trading Computation for Communication: Distributed Stochastic Dual Coordinate Ascent , 2013, NIPS.
[7] Jakub Marecek,et al. Inequality-Constrained Matrix Completion: Adding the Obvious Helps! , 2014, ArXiv.
[8] John N. Tsitsiklis,et al. Distributed Asynchronous Deterministic and Stochastic Gradient Optimization Algorithms , 1984, 1984 American Control Conference.
[9] Peter Richtárik,et al. Parallel coordinate descent methods for big data optimization , 2012, Mathematical Programming.
[10] Ohad Shamir,et al. Communication-Efficient Distributed Optimization using an Approximate Newton-type Method , 2013, ICML.
[11] Gideon S. Mann,et al. Efficient Large-Scale Distributed Training of Conditional Maximum Entropy Models , 2009, NIPS.
[12] LinChih-Jen,et al. Large Linear Classification When Data Cannot Fit in Memory , 2012 .
[13] Maria-Florina Balcan,et al. Distributed Learning, Communication Complexity and Privacy , 2012, COLT.
[14] Chih-Jen Lin,et al. A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.
[15] Chih-Jen Lin,et al. Large Linear Classification When Data Cannot Fit in Memory , 2011, TKDD.
[16] Alexander J. Smola,et al. Parallelized Stochastic Gradient Descent , 2010, NIPS.
[17] Avleen Singh Bijral,et al. Mini-Batch Primal and Dual Methods for SVMs , 2013, ICML.
[18] Michael J. Franklin,et al. Resilient Distributed Datasets: A Fault-Tolerant Abstraction for In-Memory Cluster Computing , 2012, NSDI.
[19] Ohad Shamir,et al. Optimal Distributed Online Prediction , 2011, ICML.
[20] Rong Jin,et al. Analysis of Distributed Stochastic Dual Coordinate Ascent , 2013, 1312.1031.
[21] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[22] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[23] Léon Bottou,et al. Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.
[24] Joseph K. Bradley,et al. Parallel Coordinate Descent for L1-Regularized Loss Minimization , 2011, ICML.
[25] Peter Richtárik,et al. Distributed Block Coordinate Descent for Minimizing Partially Separable Functions , 2014, 1406.0238.
[26] Michael I. Jordan,et al. Estimation, Optimization, and Parallelism when Data is Sparse , 2013, NIPS.
[27] Ion Necoara,et al. Efficient parallel coordinate descent algorithm for convex optimization problems with separable constraints: Application to distributed MPC , 2013, 1302.3092.
[28] Rong Jin,et al. On Theoretical Analysis of Distributed Stochastic Dual Coordinate Ascent , 2013, ArXiv.
[29] Peter Richtárik,et al. Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function , 2011, Mathematical Programming.
[30] Ohad Shamir,et al. Optimal Distributed Online Prediction Using Mini-Batches , 2010, J. Mach. Learn. Res..
[31] Yoram Singer,et al. Pegasos: primal estimated sub-gradient solver for SVM , 2011, Math. Program..
[32] H. Robbins. A Stochastic Approximation Method , 1951 .