DeceFL: A Principled Decentralized Federated Learning Framework

Traditional machine learning relies on a centralized data pipeline, i.e., data are provided to a central server for model training. In many applications, however, data are inherently fragmented. Such a decentralized nature of these databases presents the biggest challenge for collaboration: sending all decentralized datasets to a central server raises serious privacy concerns. Although there has been a joint effort in tackling such a critical issue by proposing privacy-preserving machine learning frameworks, such as federated learning, most state-of-the-art frameworks are built still in a centralized way, in which a central client is needed for collecting and distributing model information (instead of data itself) from every other client, leading to high communication pressure and high vulnerability when there exists a failure at or attack on the central client. Here we propose a principled decentralized federated learning algorithm (DeceFL), which does not require a central client and relies only on local information transmission between clients and their neighbors, representing a fully decentralized learning framework. It has been further proven that every client reaches the global minimum with zero performance gap and achieves the same convergence rate O(1/T) (where T is the number of iterations in gradient descent) as centralized federated learning when the loss function is smooth and strongly convex. Finally, the proposed algorithm has been applied to a number of applications to illustrate its effectiveness for both convex and nonconvex loss functions, demonstrating its applicability to a wide range of real-world medical and industrial applications. 1School of Artificial Intelligence and Automation, Huazhong University of Science and Technology. 2School of Mechanical Science and Engineering, Huazhong University of Science and Technology. 3Department of Applied Mathematics, University of Waterloo. 4State Key Laboratory of Synthetical Automation for Process Industries, Northeastern University. 5School of Electrical Engineering and Computer Science, and Digital Futures, KTH Royal Institute of Technology. 6AVIC Chengdu Aircraft Industrial (Group) Co., Ltd.. ∗Equal contributions. Email: yye@hust.edu.cn. November 2, 2021 DRAFT ar X iv :2 10 7. 07 17 1v 2 [ cs .L G ] 3 0 O ct 2 02 1

[1]  Hubert Eichner,et al.  Towards Federated Learning at Scale: System Design , 2019, SysML.

[2]  P. Lancaster,et al.  The theory of matrices : with applications , 1985 .

[3]  Ziyang Meng,et al.  A survey of distributed optimization , 2019, Annu. Rev. Control..

[4]  Andreas Keller,et al.  Swarm Learning for decentralized and confidential clinical machine learning , 2021, Nature.

[5]  Tianjian Chen,et al.  Federated Machine Learning: Concept and Applications , 2019 .

[6]  Xiang Li,et al.  On the Convergence of FedAvg on Non-IID Data , 2019, ICLR.

[7]  Wei Shi,et al.  Achieving Geometric Convergence for Distributed Optimization Over Time-Varying Graphs , 2016, SIAM J. Optim..

[8]  Andrew Chi-Chih Yao,et al.  Protocols for secure computations , 1982, FOCS 1982.

[9]  Yu Yu,et al.  Privacy-preserving computation in the post-quantum era , 2021, National science review.

[10]  Christian Etmann,et al.  Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans , 2020 .

[11]  Michael O. Rabin,et al.  In Foundations of secure computation , 1978 .

[12]  Richard Nock,et al.  Advances and Open Problems in Federated Learning , 2021, Found. Trends Mach. Learn..

[13]  Joseph D. Janizek,et al.  AI for radiographic COVID-19 detection selects shortcuts over signal , 2020, Nature Machine Intelligence.

[14]  Shaohuai Shi,et al.  Communication-Efficient Decentralized Learning with Sparsification and Adaptive Peer Selection , 2020, 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS).

[15]  Qing Ling,et al.  EXTRA: An Exact First-Order Algorithm for Decentralized Consensus Optimization , 2014, 1404.6264.

[16]  Richard M. Murray,et al.  Consensus problems in networks of agents with switching topology and time-delays , 2004, IEEE Transactions on Automatic Control.

[17]  Kin K. Leung,et al.  Adaptive Federated Learning in Resource Constrained Edge Computing Systems , 2018, IEEE Journal on Selected Areas in Communications.

[18]  L. Mombaerts,et al.  An interpretable mortality prediction model for COVID-19 patients , 2020, Nature Machine Intelligence.

[19]  Rachid Guerraoui,et al.  Personalized and Private Peer-to-Peer Machine Learning , 2017, AISTATS.

[20]  Wei Zhang,et al.  Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent , 2017, NIPS.

[21]  Martin Jaggi,et al.  Decentralized Stochastic Optimization and Gossip Algorithms with Compressed Communication , 2019, ICML.

[22]  Randal W. Beard,et al.  Consensus seeking in multiagent systems under dynamically changing interaction topologies , 2005, IEEE Transactions on Automatic Control.

[23]  W. Price,et al.  Privacy in the age of medical big data , 2019, Nature Medicine.

[24]  Jie Lin,et al.  Coordination of groups of mobile autonomous agents using nearest neighbor rules , 2003, IEEE Trans. Autom. Control..

[25]  Alan M. Frieze,et al.  Random graphs , 2006, SODA '06.

[26]  Na Li,et al.  Accelerated Distributed Nesterov Gradient Descent , 2017, IEEE Transactions on Automatic Control.

[27]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[28]  Shuo Yang,et al.  Distributed Machine Learning on Mobile Devices: A Survey , 2019, ArXiv.

[29]  Cheng Cheng,et al.  A general end-to-end diagnosis framework for manufacturing systems , 2019 .

[30]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[31]  Peter Richtárik,et al.  Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.

[32]  Han Yu,et al.  Threats to Federated Learning: A Survey , 2020, ArXiv.

[33]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[34]  John N. Tsitsiklis,et al.  Convergence Speed in Distributed Consensus and Averaging , 2009, SIAM J. Control. Optim..

[35]  Daniel Rueckert,et al.  End-to-end privacy preserving deep learning on multi-institutional medical imaging , 2021, Nature Machine Intelligence.