暂无分享,去创建一个
Dimitris S. Papailiopoulos | Hongyi Wang | Lingjiao Chen | Zachary B. Charles | Hongyi Wang | Dimitris Papailiopoulos | Lingjiao Chen
[1] Dimitris S. Papailiopoulos,et al. Approximate Gradient Coding via Sparse Random Graphs , 2017, ArXiv.
[2] Dimitris S. Papailiopoulos,et al. Perturbed Iterate Analysis for Asynchronous Stochastic Optimization , 2015, SIAM J. Optim..
[3] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[4] Amir Salman Avestimehr,et al. Coded Computation Over Heterogeneous Clusters , 2019, IEEE Transactions on Information Theory.
[5] Martin J. Wainwright,et al. Distributed Dual Averaging In Networks , 2010, NIPS.
[6] Lisandro Dalcin,et al. Parallel distributed computing using Python , 2011 .
[7] Rachid Guerraoui,et al. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent , 2017, NIPS.
[8] Robert S. Boyer,et al. MJRTY: A Fast Majority Vote Algorithm , 1991, Automated Reasoning: Essays in Honor of Woody Bledsoe.
[9] Pulkit Grover,et al. Coded convolution for parallel and distributed computing within a deadline , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).
[10] Yann LeCun,et al. Deep learning with Elastic Averaging SGD , 2014, NIPS.
[11] Leslie Lamport,et al. The Byzantine Generals Problem , 1982, TOPL.
[12] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[13] Sarvar Patel,et al. Practical Secure Aggregation for Federated Learning on User-Held Data , 2016, ArXiv.
[14] Miguel Oom Temudo de Castro,et al. Practical Byzantine fault tolerance , 1999, OSDI '99.
[15] Zheng Zhang,et al. MXNet: A Flexible and Efficient Machine Learning Library for Heterogeneous Distributed Systems , 2015, ArXiv.
[16] Trishul M. Chilimbi,et al. Project Adam: Building an Efficient and Scalable Deep Learning Training System , 2014, OSDI.
[17] Tong Zhang,et al. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction , 2013, NIPS.
[18] Alexandros G. Dimakis,et al. Gradient Coding: Avoiding Stragglers in Distributed Learning , 2017, ICML.
[19] Marc'Aurelio Ranzato,et al. Large Scale Distributed Deep Networks , 2012, NIPS.
[20] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[21] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[22] Ramakrishna Kotla,et al. Zyzzyva: speculative byzantine fault tolerance , 2007, TOCS.
[23] Randy H. Katz,et al. Improving MapReduce Performance in Heterogeneous Environments , 2008, OSDI.
[24] Jakub Konecný,et al. Federated Optimization: Distributed Optimization Beyond the Datacenter , 2015, ArXiv.
[25] Samy Bengio,et al. Revisiting Distributed Synchronous SGD , 2016, ArXiv.
[26] Lili Su,et al. Distributed Statistical Machine Learning in Adversarial Settings: Byzantine Gradient Descent , 2017, Abstracts of the 2018 ACM International Conference on Measurement and Modeling of Computer Systems.
[27] Alexander J. Smola,et al. Scaling Distributed Machine Learning with the Parameter Server , 2014, OSDI.
[28] Ohad Shamir,et al. Better Mini-Batch Algorithms via Accelerated Gradient Methods , 2011, NIPS.
[29] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[30] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[31] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[32] Stephen J. Wright,et al. Hogwild: A Lock-Free Approach to Parallelizing Stochastic Gradient Descent , 2011, NIPS.
[33] Dan Alistarh,et al. QSGD: Communication-Optimal Stochastic Gradient Descent, with Applications to Training Neural Networks , 2016, 1610.02132.
[34] Bo Pang,et al. Seeing Stars: Exploiting Class Relationships for Sentiment Categorization with Respect to Rating Scales , 2005, ACL.
[35] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Stephen J. Wright,et al. An asynchronous parallel stochastic coordinate descent algorithm , 2013, J. Mach. Learn. Res..
[37] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[38] Nicholas Pippenger,et al. Reliable computation by formulas in the presence of noise , 1988, IEEE Trans. Inf. Theory.
[39] Alexandros G. Dimakis,et al. Gradient Coding From Cyclic MDS Codes and Expander Graphs , 2017, IEEE Transactions on Information Theory.
[40] J. Neumann. Probabilistic Logic and the Synthesis of Reliable Organisms from Unreliable Components , 1956 .
[41] Dimitris S. Papailiopoulos,et al. Speeding up distributed machine learning using codes , 2016, ISIT.
[42] Soummya Kar,et al. Coded Distributed Computing for Inverse Problems , 2017, NIPS.
[43] Thomas Hofmann,et al. Communication-Efficient Distributed Dual Coordinate Ascent , 2014, NIPS.
[44] Pulkit Grover,et al. “Short-Dot”: Computing Large Linear Transforms Distributedly Using Coded Short Dot Products , 2017, IEEE Transactions on Information Theory.