Distributed Asynchronous Dual-Free Stochastic Dual Coordinate Ascent
暂无分享,去创建一个
[1] Thomas Hofmann,et al. Communication-Efficient Distributed Dual Coordinate Ascent , 2014, NIPS.
[2] Shai Shalev-Shwartz,et al. Accelerated Mini-Batch Stochastic Dual Coordinate Ascent , 2013, NIPS.
[3] Alexander J. Smola,et al. Communication Efficient Distributed Machine Learning with the Parameter Server , 2014, NIPS.
[4] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[5] Zeyuan Allen Zhu,et al. Improved SVRG for Non-Strongly-Convex or Sum-of-Non-Convex Objectives , 2015, ICML.
[6] John Langford,et al. Slow Learners are Fast , 2009, NIPS.
[7] James T. Kwok,et al. Asynchronous Distributed Semi-Stochastic Gradient Optimization , 2015, AAAI.
[8] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[9] Peter Richtárik,et al. Stochastic Dual Coordinate Ascent with Adaptive Probabilities , 2015, ICML.
[10] James T. Kwok,et al. Asynchronous Distributed ADMM for Consensus Optimization , 2014, ICML.
[11] Shai Shalev-Shwartz,et al. SDCA without Duality, Regularization, and Individual Convexity , 2016, ICML.
[12] Michael I. Jordan,et al. Adding vs. Averaging in Distributed Primal-Dual Optimization , 2015, ICML.
[13] Léon Bottou,et al. Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.
[14] Alexander J. Smola,et al. On Variance Reduction in Stochastic Gradient Descent and its Asynchronous Variants , 2015, NIPS.
[15] James T. Kwok,et al. Fast Distributed Asynchronous SGD with Variance Reduction , 2015, ArXiv.
[16] Mingyi Hong,et al. A Distributed, Asynchronous, and Incremental Algorithm for Nonconvex Optimization: An ADMM Approach , 2014, IEEE Transactions on Control of Network Systems.
[17] Lin Xiao,et al. A Proximal Stochastic Gradient Method with Progressive Variance Reduction , 2014, SIAM J. Optim..
[18] Heng Huang,et al. Asynchronous Stochastic Gradient Descent with Variance Reduction for Non-Convex Optimization , 2016, AAAI 2016.
[19] Chih-Jen Lin,et al. A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.
[20] Alexander J. Smola,et al. Efficient mini-batch training for stochastic optimization , 2014, KDD.
[21] Carlos Guestrin,et al. Distributed GraphLab : A Framework for Machine Learning and Data Mining in the Cloud , 2012 .
[22] Elad Hazan,et al. Fast and Simple PCA via Convex Optimization , 2015, ArXiv.
[23] Yijun Huang,et al. Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization , 2015, NIPS.
[24] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[25] George Bosilca,et al. Open MPI: Goals, Concept, and Design of a Next Generation MPI Implementation , 2004, PVM/MPI.
[26] Shai Shalev-Shwartz,et al. SDCA without Duality , 2015, ArXiv.
[27] Xi He,et al. Dual Free SDCA for Empirical Risk Minimization with Adaptive Probabilities , 2015, ArXiv.
[28] Tong Zhang,et al. Stochastic Optimization with Importance Sampling for Regularized Loss Minimization , 2014, ICML.
[29] Alexander J. Smola,et al. Fast Stochastic Methods for Nonsmooth Nonconvex Optimization , 2016, ArXiv.
[30] Wu-Jun Li,et al. Fast Asynchronous Parallel Stochastic Gradient Descent: A Lock-Free Approach with Convergence Guarantee , 2016, AAAI.
[31] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[32] Mark W. Schmidt,et al. Minimizing finite sums with the stochastic average gradient , 2013, Mathematical Programming.
[33] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[34] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[35] John C. Duchi,et al. Distributed delayed stochastic optimization , 2011, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[36] Tianbao Yang,et al. Trading Computation for Communication: Distributed Stochastic Dual Coordinate Ascent , 2013, NIPS.
[37] Augustus Odena,et al. Faster Asynchronous SGD , 2016, ArXiv.
[38] Ji Liu,et al. Staleness-Aware Async-SGD for Distributed Deep Learning , 2015, IJCAI.
[39] Stephen J. Wright,et al. An Asynchronous Parallel Randomized Kaczmarz Algorithm , 2014, ArXiv.
[40] Peter Richtárik,et al. Distributed Mini-Batch SDCA , 2015, ArXiv.
[41] Avleen Singh Bijral,et al. Mini-Batch Primal and Dual Methods for SVMs , 2013, ICML.