Multi-Stage Hybrid Federated Learning over Large-Scale Wireless Fog Networks

One of the popular methods for distributed machine learning (ML) is federated learning, in which devices train local models based on their datasets, which are in turn aggregated periodically by a server. In large-scale fog networks, the "star" learning topology of federated learning poses several challenges in terms of resource utilization. We develop multi-stage hybrid model training (MH-MT), a novel learning methodology for distributed ML in these scenarios. Leveraging the hierarchical structure of fog systems, MH-MT combines multi-stage parameter relaying with distributed consensus formation among devices in a hybrid learning paradigm across network layers. We theoretically derive the convergence bound of MH-MT with respect to the network topology, ML model, and algorithm parameters such as the rounds of consensus employed in different clusters of devices. We obtain a set of policies for the number of consensus rounds at different clusters to guarantee either a finite optimality gap or convergence to the global optimum. Subsequently, we develop an adaptive distributed control algorithm for MH-MT to tune the number of consensus rounds at each cluster of local devices over time to meet convergence criteria. Our numerical experiments validate the performance of MH-MT in terms of convergence speed and resource utilization.

[1]  Peter Richtárik,et al.  Distributed Coordinate Descent Method for Learning with Big Data , 2013, J. Mach. Learn. Res..

[2]  Sherali Zeadally,et al.  Vehicular ad hoc networks (VANETS): status, results, and challenges , 2010, Telecommunication Systems.

[3]  Dan Alistarh,et al.  SparCML: high-performance sparse communication for machine learning , 2018, SC.

[4]  Dipankar Raychaudhuri,et al.  Scalability and Performance Evaluation of Edge Cloud Systems for Latency Constrained Applications , 2018, 2018 IEEE/ACM Symposium on Edge Computing (SEC).

[5]  Solmaz Niknam,et al.  Federated Learning for Wireless Communications: Motivation, Opportunities, and Challenges , 2019, IEEE Communications Magazine.

[6]  Richard Nock,et al.  Private federated learning on vertically partitioned data via entity resolution and additively homomorphic encryption , 2017, ArXiv.

[7]  Soummya Kar,et al.  Distributed Consensus Algorithms in Sensor Networks With Imperfect Communication: Link Failures and Channel Noise , 2007, IEEE Transactions on Signal Processing.

[8]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[9]  Guo-Xing Wen,et al.  Adaptive Consensus Control for a Class of Nonlinear Multiagent Time-Delay Systems Using Neural Networks , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[10]  Omid Semiari,et al.  Federated Learning in the Sky: Joint Power Allocation and Scheduling with UAV Swarms , 2020, ICC 2020 - 2020 IEEE International Conference on Communications (ICC).

[11]  Xue Li,et al.  Dynamic Sampling and Selective Masking for Communication-Efficient Federated Learning , 2020, ArXiv.

[12]  Debarshi Kumar Sanyal,et al.  An overview of device-to-device communication in cellular networks , 2017, ICT Express.

[13]  Anna Scaglione,et al.  Distributed Constrained Optimization by Consensus-Based Primal-Dual Perturbation Method , 2013, IEEE Transactions on Automatic Control.

[14]  Monica Nicoli,et al.  Federated Learning With Cooperating Devices: A Consensus Approach for Massive IoT Networks , 2019, IEEE Internet of Things Journal.

[15]  Huaiyu Dai,et al.  From Federated Learning to Fog Learning: Towards Large-Scale Distributed Machine Learning in Heterogeneous Wireless Networks , 2020, ArXiv.

[16]  Ilker Bekmezci,et al.  Flying Ad-Hoc Networks (FANETs): A survey , 2013, Ad Hoc Networks.

[17]  Canh Dinh,et al.  Federated Learning Over Wireless Networks: Convergence Analysis and Resource Allocation , 2019, IEEE/ACM Transactions on Networking.

[18]  Jason Weston,et al.  A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.

[19]  Mehdi Bennis,et al.  GADMM: Fast and Communication Efficient Framework for Distributed Machine Learning , 2019, J. Mach. Learn. Res..

[20]  Eryk Dutkiewicz,et al.  A review of routing protocols for mobile ad hoc networks , 2004, Ad Hoc Networks.

[21]  Halim Yanikomeroglu,et al.  Device-to-device communication in 5G cellular networks: challenges, solutions, and future directions , 2014, IEEE Communications Magazine.

[22]  Stephen P. Boyd,et al.  Fast linear iterations for distributed averaging , 2003, 42nd IEEE International Conference on Decision and Control (IEEE Cat. No.03CH37475).

[23]  Jingyan Jiang,et al.  Decentralized Federated Learning: A Segmented Gossip Approach , 2019, ArXiv.

[24]  Kin K. Leung,et al.  Adaptive Federated Learning in Resource Constrained Edge Computing Systems , 2018, IEEE Journal on Selected Areas in Communications.

[25]  Walid Saad,et al.  A Joint Learning and Communications Framework for Federated Learning Over Wireless Networks , 2021, IEEE Transactions on Wireless Communications.

[26]  Kaibin Huang,et al.  Towards an Intelligent Edge: Wireless Communication Meets Machine Learning , 2018, ArXiv.

[27]  Carlee Joe-Wong,et al.  Network-Aware Optimization of Distributed Learning for Fog Computing , 2020, IEEE/ACM Transactions on Networking.

[28]  Mohamed F. Younis,et al.  A survey on routing protocols for wireless sensor networks , 2005, Ad Hoc Networks.

[29]  Aryan Mokhtari,et al.  FedPAQ: A Communication-Efficient Federated Learning Method with Periodic Averaging and Quantization , 2019, AISTATS.

[30]  Tao Li,et al.  Consensus Conditions of Multi-Agent Systems With Time-Varying Topologies and Stochastic Communication Noises , 2010, IEEE Transactions on Automatic Control.

[31]  Karl Henrik Johansson,et al.  Subgradient methods and consensus algorithms for solving convex optimization problems , 2008, 2008 47th IEEE Conference on Decision and Control.

[32]  Tassilo Klein,et al.  Differentially Private Federated Learning: A Client Level Perspective , 2017, ArXiv.

[33]  Peter Richtárik,et al.  Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.

[34]  Miao Pan,et al.  Federated Learning in Vehicular Edge Computing: A Selective Model Aggregation Approach , 2020, IEEE Access.