暂无分享,去创建一个
Tong Zhang | Wei Xu | Fen Xia | Jialei Wang | Shun Zheng | Jialei Wang | Tong Zhang | W. Xu | Fen Xia | Shun Zheng
[1] Michael I. Jordan,et al. Adding vs. Averaging in Distributed Primal-Dual Optimization , 2015, ICML.
[2] Sham M. Kakade,et al. Un-regularizing: approximate proximal point and faster stochastic algorithms for empirical risk minimization , 2015, ICML.
[3] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[4] S. Sundararajan,et al. A distributed block coordinate descent method for training $l_1$ regularized linear classifiers , 2014, J. Mach. Learn. Res..
[5] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[6] Ohad Shamir,et al. Communication-Efficient Distributed Optimization using an Approximate Newton-type Method , 2013, ICML.
[7] Yuchen Zhang,et al. Stochastic Primal-Dual Coordinate Method for Regularized Empirical Risk Minimization , 2014, ICML.
[8] Rong Jin,et al. Analysis of Distributed Stochastic Dual Coordinate Ascent , 2013, 1312.1031.
[9] R. Tibshirani,et al. A note on the group lasso and a sparse group lasso , 2010, 1001.0736.
[10] S. Sathiya Keerthi,et al. An efficient distributed learning algorithm based on effective local functional approximations , 2018, J. Mach. Learn. Res..
[11] Yurii Nesterov,et al. Smooth minimization of non-smooth functions , 2005, Math. Program..
[12] Dan Roth,et al. Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM , 2015, ICML.
[13] Stephen P. Boyd,et al. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..
[14] Alexander J. Smola,et al. Parallelized Stochastic Gradient Descent , 2010, NIPS.
[15] Francis Bach,et al. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives , 2014, NIPS.
[16] Peter Richtárik,et al. Distributed Mini-Batch SDCA , 2015, ArXiv.
[17] Michael I. Jordan,et al. Distributed optimization with arbitrary local solvers , 2015, Optim. Methods Softw..
[18] Tianbao Yang,et al. Trading Computation for Communication: Distributed Stochastic Dual Coordinate Ascent , 2013, NIPS.
[19] Michael I. Jordan,et al. CoCoA: A General Framework for Communication-Efficient Distributed Optimization , 2016, J. Mach. Learn. Res..
[20] Inderjit S. Dhillon,et al. PASSCoDe: Parallel ASynchronous Stochastic dual Co-ordinate Descent , 2015, ICML.
[21] Shai Shalev-Shwartz,et al. Stochastic dual coordinate ascent methods for regularized loss , 2012, J. Mach. Learn. Res..
[22] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[23] Chih-Jen Lin,et al. A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.
[24] Avleen Singh Bijral,et al. Mini-Batch Primal and Dual Methods for SVMs , 2013, ICML.
[25] D K Smith,et al. Numerical Optimization , 2001, J. Oper. Res. Soc..
[26] Jorge Nocedal,et al. On the limited memory BFGS method for large scale optimization , 1989, Math. Program..
[27] Richard L. Graham,et al. Open MPI: A Flexible High Performance MPI , 2005, PPAM.
[28] Jianfeng Gao,et al. Scalable training of L1-regularized log-linear models , 2007, ICML '07.
[29] Alexander J. Smola,et al. Scaling Distributed Machine Learning with the Parameter Server , 2014, OSDI.
[30] Yuchen Zhang,et al. DiSCO: Distributed Optimization for Self-Concordant Empirical Loss , 2015, ICML.
[31] Sanjay Ghemawat,et al. MapReduce: Simplified Data Processing on Large Clusters , 2004, OSDI.
[32] Y. Nesterov. A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .
[33] Tong Zhang,et al. Accelerated proximal stochastic dual coordinate ascent for regularized loss minimization , 2013, Mathematical Programming.
[34] Emmanuel J. Candès,et al. Adaptive Restart for Accelerated Gradient Schemes , 2012, Foundations of Computational Mathematics.
[35] Michael J. Franklin,et al. Resilient Distributed Datasets: A Fault-Tolerant Abstraction for In-Memory Cluster Computing , 2012, NSDI.
[36] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[37] Thomas Hofmann,et al. Communication-Efficient Distributed Dual Coordinate Ascent , 2014, NIPS.
[38] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.