暂无分享,去创建一个
Aryan Mokhtari | Ramtin Pedarsani | Hamed Hassani | Amirhossein Reisizadeh | Isidoros Tziotis | Aryan Mokhtari | Hamed Hassani | Ramtin Pedarsani | Isidoros Tziotis | Amirhossein Reisizadeh
[1] Aryan Mokhtari,et al. First-Order Adaptive Sample Size Methods to Reduce Complexity of Empirical Risk Minimization , 2017, NIPS.
[2] O. Bousquet. Concentration Inequalities and Empirical Processes Theory Applied to the Analysis of Learning Algorithms , 2002 .
[3] Kannan Ramchandran,et al. Speeding Up Distributed Machine Learning Using Codes , 2015, IEEE Transactions on Information Theory.
[4] Aryan Mokhtari,et al. Efficient Distributed Hessian Free Algorithm for Large-scale Empirical Risk Minimization via Accumulating Sample Strategy , 2018, AISTATS.
[5] Manzil Zaheer,et al. Adaptive Federated Optimization , 2020, ICLR.
[6] Takayuki Nishio,et al. Client Selection for Federated Learning with Heterogeneous Resources in Mobile Edge , 2018, ICC 2019 - 2019 IEEE International Conference on Communications (ICC).
[7] Konstantin Mishchenko,et al. Tighter Theory for Local SGD on Identical and Heterogeneous Data , 2020, AISTATS.
[8] Peter Richtárik,et al. Distributed Learning with Compressed Gradient Differences , 2019, ArXiv.
[9] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[10] Anit Kumar Sahu,et al. Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.
[11] Ali Jadbabaie,et al. Robust Federated Learning: The Case of Affine Distribution Shifts , 2020, NeurIPS.
[12] Sham M. Kakade,et al. Competing with the Empirical Risk Minimizer in a Single Pass , 2014, COLT.
[13] Sebastian U. Stich,et al. The Error-Feedback Framework: Better Rates for SGD with Delayed Gradients and Compressed Communication , 2019, ArXiv.
[14] Aryan Mokhtari,et al. Federated Learning with Compression: Unified Analysis and Sharp Guarantees , 2020, AISTATS.
[15] Amir Salman Avestimehr,et al. Coded computation over heterogeneous clusters , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).
[16] Farzin Haddadpour,et al. On the Convergence of Local Descent Methods in Federated Learning , 2019, ArXiv.
[17] Aryan Mokhtari,et al. FedPAQ: A Communication-Efficient Federated Learning Method with Periodic Averaging and Quantization , 2019, AISTATS.
[18] Farzin Haddadpour,et al. Local SGD with Periodic Averaging: Tighter Analysis and Adaptive Synchronization , 2019, NeurIPS.
[19] Aryan Mokhtari,et al. Adaptive Newton Method for Empirical Risk Minimization to Statistical Accuracy , 2016, NIPS.
[20] Mehryar Mohri,et al. Agnostic Federated Learning , 2019, ICML.
[21] Sebastian U. Stich,et al. Local SGD Converges Fast and Communicates Little , 2018, ICLR.
[22] Aryan Mokhtari,et al. Efficient Nonconvex Empirical Risk Minimization via Adaptive Sample Size Methods , 2019, AISTATS.
[23] Kin K. Leung,et al. Adaptive Federated Learning in Resource Constrained Edge Computing Systems , 2018, IEEE Journal on Selected Areas in Communications.
[24] Qinghua Liu,et al. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization , 2020, NeurIPS.
[25] Vladimir N. Vapnik,et al. The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.
[26] Indranil Gupta,et al. Asynchronous Federated Optimization , 2019, ArXiv.
[27] Enhong Chen,et al. Variance Reduced Local SGD with Lower Communication Complexity , 2019, ArXiv.
[28] Laurent Condat,et al. From Local SGD to Local Fixed Point Methods for Federated Learning , 2020, ICML.
[29] Fan Zhou,et al. On the convergence properties of a K-step averaging stochastic gradient descent algorithm for nonconvex optimization , 2017, IJCAI.
[30] Jianyu Wang,et al. Cooperative SGD: A unified Framework for the Design and Analysis of Communication-Efficient SGD Algorithms , 2018, ArXiv.
[31] Jianyu Wang,et al. Adaptive Communication Strategies to Achieve the Best Error-Runtime Trade-off in Local-Update SGD , 2018, MLSys.
[32] Aryan Mokhtari,et al. Large Scale Empirical Risk Minimization via Truncated Adaptive Newton Method , 2017, AISTATS.
[33] Xiang Li,et al. On the Convergence of FedAvg on Non-IID Data , 2019, ICLR.
[34] Mehryar Mohri,et al. SCAFFOLD: Stochastic Controlled Averaging for On-Device Federated Learning , 2019, ArXiv.
[35] Anit Kumar Sahu,et al. Federated Optimization in Heterogeneous Networks , 2018, MLSys.
[36] Martin Jaggi,et al. A Unified Theory of Decentralized SGD with Changing Topology and Local Updates , 2020, ICML.
[37] Martin J. Wainwright,et al. FedSplit: An algorithmic framework for fast federated optimization , 2020, NeurIPS.
[38] Lawrence Carin,et al. Faster On-Device Training Using New Federated Momentum Algorithm , 2020, ArXiv.
[39] Michael I. Jordan,et al. Convexity, Classification, and Risk Bounds , 2006 .
[40] Richard Nock,et al. Advances and Open Problems in Federated Learning , 2019, Found. Trends Mach. Learn..