Distributed Algorithms for Composite Optimization: Unified and Tight Convergence Analysis

We study distributed composite optimization over networks: agents minimize a sum of smooth (strongly) convex functions–the agents’ sum-utility–plus a nonsmooth (extendedvalued) convex one. We propose a general unified algorithmic framework for such a class of problems and provide a unified convergence analysis leveraging the theory of operator splitting. Distinguishing features of our scheme are: (i) When the agents’ functions are strongly convex, the algorithm converges at a linear rate, whose dependence on the agents’ functions and network topology is decoupled, matching the typical rates of centralized optimization; the rate expression is sharp and improves on existing results; (ii) When the objective function is convex (but not strongly convex), similar separation as in (i) is established for the coefficient of the proved sublinear rate; (iii) The algorithm can adjust the ratio between the number of communications and computations to achieve a rate (in terms of computations) independent on the network connectivity; and (iv) A by-product of our analysis is a tuning recommendation for several existing (non accelerated) distributed algorithms yielding the fastest provably (worst-case) convergence rate. This is the first time that a general distributed algorithmic framework applicable to composite optimization enjoys all such properties.

[1]  Lihua Xie,et al.  Convergence of Asynchronous Distributed Gradient Methods Over Stochastic Networks , 2018, IEEE Transactions on Automatic Control.

[2]  Ermin Wei,et al.  A General Framework of Exact Primal-Dual First-Order Algorithms for Distributed Optimization ∗ , 2019, 2019 IEEE 58th Conference on Decision and Control (CDC).

[3]  Laurent Massoulié,et al.  Optimal Algorithms for Smooth and Strongly Convex Distributed Optimization in Networks , 2017, ICML.

[4]  Stephen P. Boyd,et al.  Fast linear iterations for distributed averaging , 2003, 42nd IEEE International Conference on Decision and Control (IEEE Cat. No.03CH37475).

[5]  Lihua Xie,et al.  A Bregman Splitting Scheme for Distributed Optimization Over Networks , 2018, IEEE Transactions on Automatic Control.

[6]  Bryan Van Scoy,et al.  A Canonical Form for First-Order Distributed Optimization Algorithms , 2018, 2019 American Control Conference (ACC).

[7]  Martin J. Wainwright,et al.  Fast global convergence rates of gradient methods for high-dimensional statistical recovery , 2010, NIPS.

[8]  Ali H. Sayed,et al.  Exact Diffusion for Distributed Optimization and Learning—Part II: Convergence Analysis , 2017, IEEE Transactions on Signal Processing.

[9]  N. S. Aybat,et al.  Distributed Linearized Alternating Direction Method of Multipliers for Composite Convex Consensus Optimization , 2015, IEEE Transactions on Automatic Control.

[10]  Qing Ling,et al.  A Proximal Gradient Algorithm for Decentralized Composite Optimization , 2015, IEEE Transactions on Signal Processing.

[11]  Lihua Xie,et al.  Augmented distributed gradient methods for multi-agent optimization under uncoordinated constant stepsizes , 2015, 2015 54th IEEE Conference on Decision and Control (CDC).

[12]  Wei Shi,et al.  Achieving Geometric Convergence for Distributed Optimization Over Time-Varying Graphs , 2016, SIAM J. Optim..

[13]  Dusan Jakovetic,et al.  A Unification and Generalization of Exact Distributed First-Order Methods , 2017, IEEE Transactions on Signal and Information Processing over Networks.

[14]  Na Li,et al.  Harnessing smoothness to accelerate distributed optimization , 2016, 2016 IEEE 55th Conference on Decision and Control (CDC).

[15]  Ying Sun,et al.  A Unified Contraction Analysis of a Class of Distributed Algorithms for Composite Optimization , 2019, 2019 IEEE 8th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP).

[16]  Ali H. Sayed,et al.  A Linearly Convergent Proximal Gradient Algorithm for Decentralized Optimization , 2019, NeurIPS.

[17]  Ying Sun,et al.  Convergence Rate of Distributed Optimization Algorithms Based on Gradient Tracking , 2019, ArXiv.

[18]  Sébastien Bubeck,et al.  Convex Optimization: Algorithms and Complexity , 2014, Found. Trends Mach. Learn..

[19]  Ying Sun,et al.  Accelerated Primal-Dual Algorithms for Distributed Smooth Convex Optimization over Networks , 2020, AISTATS.

[20]  Gesualdo Scutari,et al.  Distributed nonconvex constrained optimization over time-varying digraphs , 2018, Mathematical Programming.

[21]  Wei Shi,et al.  Geometrically convergent distributed optimization with uncoordinated step-sizes , 2016, 2017 American Control Conference (ACC).

[22]  Bin Hu,et al.  Robust convergence analysis of distributed optimization algorithms , 2017, 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[23]  Bryan Van Scoy,et al.  Distributed Optimization of Nonconvex Functions over Time-Varying Graphs , 2019 .

[24]  Wei Shi,et al.  A Decentralized Proximal-Gradient Method With Network Independent Step-Sizes and Separated Convergence Rates , 2017, IEEE Transactions on Signal Processing.

[25]  Louis A. Hageman,et al.  Iterative Solution of Large Linear Systems. , 1971 .

[26]  Asuman E. Ozdaglar,et al.  Distributed Alternating Direction Method of Multipliers , 2012, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).

[27]  Gesualdo Scutari,et al.  NEXT: In-Network Nonconvex Optimization , 2016, IEEE Transactions on Signal and Information Processing over Networks.

[28]  Ali H. Sayed,et al.  Exact Diffusion for Distributed Optimization and Learning—Part I: Algorithm Development , 2017, IEEE Transactions on Signal Processing.

[29]  Qing Ling,et al.  EXTRA: An Exact First-Order Algorithm for Decentralized Consensus Optimization , 2014, 1404.6264.

[30]  Ali H. Sayed,et al.  Decentralized Proximal Gradient Algorithms with Linear Convergence Rates , 2019 .