暂无分享,去创建一个
[1] Qing Ling,et al. Communication-Censored ADMM for Decentralized Consensus Optimization , 2019, IEEE Transactions on Signal Processing.
[2] Léon Bottou,et al. Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.
[3] Michael G. Rabbat,et al. Network Topology and Communication-Computation Tradeoffs in Decentralized Optimization , 2017, Proceedings of the IEEE.
[4] Danna Zhou,et al. d. , 1840, Microbial pathogenesis.
[5] Rong Jin,et al. On the Computation and Communication Complexity of Parallel SGD with Dynamic Batch Sizes for Stochastic Non-Convex Optimization , 2019, ICML.
[6] Mark W. Schmidt,et al. Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition , 2016, ECML/PKDD.
[7] Albert S. Berahas,et al. Balancing Communication and Computation in Distributed Optimization , 2017, IEEE Transactions on Automatic Control.
[8] Dan Alistarh,et al. The Convergence of Sparsified Gradient Methods , 2018, NeurIPS.
[9] Kenneth Heafield,et al. Sparse Communication for Distributed Gradient Descent , 2017, EMNLP.
[10] G. G. Stokes. "J." , 1890, The New Yale Book of Quotations.
[11] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[12] Ameet Talwalkar,et al. Federated Multi-Task Learning , 2017, NIPS.
[13] Tsuyoshi Murata,et al. {m , 1934, ACML.
[14] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[15] Martin Jaggi,et al. Sparsified SGD with Memory , 2018, NeurIPS.
[16] Qing Ling,et al. Decentralized learning for wireless communications and networking , 2015, ArXiv.
[17] Jorge Nocedal,et al. Optimization Methods for Large-Scale Machine Learning , 2016, SIAM Rev..
[18] R. Durrett. Probability: Theory and Examples , 1993 .
[19] Hanlin Tang,et al. Decentralization Meets Quantization , 2018, ArXiv.
[20] Tianbao Yang,et al. Trading Computation for Communication: Distributed Stochastic Dual Coordinate Ascent , 2013, NIPS.
[21] Kamyar Azizzadenesheli,et al. signSGD: compressed optimisation for non-convex problems , 2018, ICML.
[22] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Yi Zhou,et al. Communication-efficient algorithms for decentralized and stochastic optimization , 2017, Mathematical Programming.
[24] Dan Alistarh,et al. QSGD: Communication-Optimal Stochastic Gradient Descent, with Applications to Training Neural Networks , 2016, 1610.02132.
[25] John B. Shoven,et al. I , Edinburgh Medical and Surgical Journal.
[26] Alexander J. Smola,et al. Communication Efficient Distributed Machine Learning with the Parameter Server , 2014, NIPS.
[27] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[28] Sebastian Caldas,et al. Expanding the Reach of Federated Learning by Reducing Client Resource Requirements , 2018, ArXiv.
[29] Martin J. Wainwright,et al. Communication-efficient algorithms for statistical optimization , 2012, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).
[30] Georgios B. Giannakis,et al. LAG: Lazily Aggregated Gradient for Communication-Efficient Distributed Learning , 2018, NeurIPS.
[31] Andrea J. Goldsmith,et al. Distributed Convex Optimization with Limited Communications , 2018, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[32] Georgios B. Giannakis,et al. Communication-Efficient Distributed Reinforcement Learning , 2018, ArXiv.
[33] Qing Ling,et al. COLA: Communication-censored Linearized ADMM for Decentralized Consensus Optimization , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).