暂无分享,去创建一个
[1] John N. Tsitsiklis,et al. Communication complexity of convex optimization , 1986, 1986 25th IEEE Conference on Decision and Control.
[2] Yurii Nesterov,et al. Cubic regularization of Newton method and its global performance , 2006, Math. Program..
[3] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[4] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[5] Martin J. Wainwright,et al. Information-theoretic lower bounds for distributed statistical estimation with communication constraints , 2013, NIPS.
[6] Alexander J. Smola,et al. Scaling Distributed Machine Learning with the Parameter Server , 2014, OSDI.
[7] Thomas Hofmann,et al. Communication-Efficient Distributed Dual Coordinate Ascent , 2014, NIPS.
[8] Ohad Shamir,et al. Communication-Efficient Distributed Optimization using an Approximate Newton-type Method , 2013, ICML.
[9] Ohad Shamir,et al. Fundamental Limits of Online and Distributed Algorithms for Statistical Learning and Estimation , 2013, NIPS.
[10] Ohad Shamir,et al. Communication Complexity of Distributed Convex Learning and Optimization , 2015, NIPS.
[11] Yuchen Zhang,et al. DiSCO: Distributed Optimization for Self-Concordant Empirical Loss , 2015, ICML.
[12] Alexander J. Smola,et al. AIDE: Fast and Communication Efficient Distributed Optimization , 2016, ArXiv.
[13] Dan Alistarh,et al. QSGD: Communication-Optimal Stochastic Gradient Descent, with Applications to Training Neural Networks , 2016, 1610.02132.
[14] Laurent Massoulié,et al. Optimal Algorithms for Smooth and Strongly Convex Distributed Optimization in Networks , 2017, ICML.
[15] Ananda Theertha Suresh,et al. Distributed Mean Estimation with Limited Communication , 2016, ICML.
[16] Yun Yang,et al. Communication-Efficient Distributed Statistical Inference , 2016, Journal of the American Statistical Association.
[17] Dan Alistarh,et al. The Convergence of Sparsified Gradient Methods , 2018, NeurIPS.
[18] Shusen Wang,et al. GIANT: Globally Improved Approximate Newton Method for Distributed Optimization , 2017, NeurIPS.
[19] Min Ye,et al. Communication-Computation Efficient Gradient Coding , 2018, ICML.
[20] Torsten Hoefler,et al. Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis. , 2018 .
[21] Sarit Khirirat,et al. Distributed learning with compressed gradients , 2018, 1806.06573.
[22] Peter Richtárik,et al. SGD and Hogwild! Convergence Without the Bounded Gradients Assumption , 2018, ICML.
[23] Daniel M. Roy,et al. NUQSGD: Improved Communication Efficiency for Data-parallel SGD via Nonuniform Quantization , 2019, ArXiv.
[24] Na Li,et al. On Maintaining Linear Convergence of Distributed Learning and Optimization Under Limited Communication , 2019, IEEE Transactions on Signal Processing.
[25] Dan Alistarh,et al. Distributed Variance Reduction with Optimal Communication , 2020 .
[26] Sindri Magnusson,et al. Communication-efficient Variance-reduced Stochastic Gradient Descent , 2020, ArXiv.
[27] Dan Alistarh,et al. Improved Communication Lower Bounds for Distributed Optimisation , 2020, ArXiv.
[28] Lin Xiao,et al. Statistically Preconditioned Accelerated Gradient Method for Distributed Optimization , 2020, ICML.
[29] Randomized Block-Diagonal Preconditioning for Parallel Learning , 2020, ICML.
[30] The Communication Complexity of Optimization , 2019, SODA.
[31] Dan Alistarh,et al. New Bounds For Distributed Mean Estimation and Variance Reduction , 2021, ICLR.
[32] Xun Qian,et al. FedNL: Making Newton-Type Methods Applicable to Federated Learning , 2021, ArXiv.
[33] Peter Richtárik,et al. Distributed Second Order Methods with Fast Rates and Compressed Communication , 2021, ICML.
[34] Tamer Basar,et al. Distributed Adaptive Newton Methods with Globally Superlinear Convergence , 2020, Autom..