On the Performance of Gradient Tracking with Local Updates
暂无分享,去创建一个
[1] Sebastian U. Stich,et al. ProxSkip: Yes! Local Gradient Steps Provably Lead to Communication Acceleration! Finally! , 2022, ICML.
[2] Sulaiman A. Alghunaim,et al. A Unified and Refined Convergence Analysis for Non-Convex Decentralized Learning , 2021, IEEE Transactions on Signal Processing.
[3] G. Scutari,et al. Distributed Optimization Based on Gradient Tracking Revisited: Enhancing Convergence Rate via Surrogation , 2020, SIAM J. Optim..
[4] Sebastian U. Stich,et al. An Improved Analysis of Gradient Tracking for Decentralized Machine Learning , 2022, NeurIPS.
[5] Wotao Yin,et al. BlueFog: Make Decentralized Algorithms Practical for Optimization and Deep Learning , 2021, ArXiv.
[6] George J. Pappas,et al. Linear Convergence in Federated Learning: Tackling Client Heterogeneity and Sparse Gradients , 2021, NeurIPS.
[7] Eduard A. Gorbunov,et al. Local SGD: Unified Theory and New Efficient Methods , 2020, AISTATS.
[8] Soummya Kar,et al. An Improved Convergence Analysis for Decentralized Online Stochastic Non-Convex Optimization , 2020, IEEE Transactions on Signal Processing.
[9] Songtao Lu,et al. Decentralized Stochastic Non-Convex Optimization over Weakly Connected Time-Varying Digraphs , 2020, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[10] Martin Jaggi,et al. A Unified Theory of Decentralized SGD with Changing Topology and Local Updates , 2020, ICML.
[11] Sashank J. Reddi,et al. SCAFFOLD: Stochastic Controlled Averaging for Federated Learning , 2019, ICML.
[12] Peter Richtárik,et al. Tighter Theory for Local SGD on Identical and Heterogeneous Data , 2019, AISTATS.
[13] Ali H. Sayed,et al. On the Influence of Bias-Correction on Distributed Stochastic Optimization , 2019, IEEE Transactions on Signal Processing.
[14] Usman A. Khan,et al. Optimization over time-varying directed graphs with row and column-stochastic matrices , 2018, 1810.07393.
[15] Tao Lin,et al. Don't Use Large Mini-Batches, Use Local SGD , 2018, ICLR.
[16] Peter Richtárik,et al. First Analysis of Local GD on Heterogeneous Data , 2019, ArXiv.
[17] Gesualdo Scutari,et al. Distributed nonconvex constrained optimization over time-varying digraphs , 2018, Mathematical Programming.
[18] Sebastian U. Stich,et al. Local SGD Converges Fast and Communicates Little , 2018, ICLR.
[19] Wei Shi,et al. A Decentralized Proximal-Gradient Method With Network Independent Step-Sizes and Separated Convergence Rates , 2017, IEEE Transactions on Signal Processing.
[20] Ali H. Sayed,et al. Exact Diffusion for Distributed Optimization and Learning—Part I: Algorithm Development , 2017, IEEE Transactions on Signal Processing.
[21] Xiangru Lian,et al. D2: Decentralized Training over Decentralized Data , 2018, ICML.
[22] Wei Shi,et al. Achieving Geometric Convergence for Distributed Optimization Over Time-Varying Graphs , 2016, SIAM J. Optim..
[23] Na Li,et al. Harnessing smoothness to accelerate distributed optimization , 2016, 2016 IEEE 55th Conference on Decision and Control (CDC).
[24] Ioannis Mitliagkas,et al. Parallel SGD: When does averaging help? , 2016, ArXiv.
[25] Gesualdo Scutari,et al. NEXT: In-Network Nonconvex Optimization , 2016, IEEE Transactions on Signal and Information Processing over Networks.
[26] Qing Ling,et al. On the Convergence of Decentralized Gradient Descent , 2013, SIAM J. Optim..
[27] Lihua Xie,et al. Augmented distributed gradient methods for multi-agent optimization under uncoordinated constant stepsizes , 2015, 2015 54th IEEE Conference on Decision and Control (CDC).
[28] Wei Shi,et al. EXTRA: An Exact First-Order Algorithm for Decentralized Consensus Optimization , 2014, SIAM J. Optim..
[29] Ali H. Sayed,et al. Distributed Pareto Optimization via Diffusion Strategies , 2012, IEEE Journal of Selected Topics in Signal Processing.
[30] Stephen P. Boyd,et al. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..
[31] Ali H. Sayed,et al. Diffusion LMS Strategies for Distributed Estimation , 2010, IEEE Transactions on Signal Processing.
[32] Angelia Nedic,et al. Distributed Stochastic Subgradient Projection Algorithms for Convex Optimization , 2008, J. Optim. Theory Appl..
[33] W. K. Hastings,et al. Monte Carlo Sampling Methods Using Markov Chains and Their Applications , 1970 .