FedFly: Toward Migration in Edge-Based Distributed Federated Learning

Federated learning (FL) is a privacy-preserving distributed machine learning technique that trains models while keeping all the original data generated on devices locally. Since devices may be resource-constrained, offloading can be used to improve FL performance by transferring computational workload from devices to edge servers. However, due to mobility, devices participating in FL may leave the network during training and need to connect to a different edge server. This is challenging because the offloaded computations from an edge server need to be migrated. In line with this assertion, we present FedFly, which is, to the best of our knowledge, the first work to migrate a deep neural network (ONN) when devices move between edge servers during FL training. Our empirical results on the CIFAR-10 dataset, with both balanced and imbalanced data distribution, support our claims that FedFly can reduce training time by up to 33 percent when a device moves after 50 percent of the training is completed, and by up to 45 percent when 90 percent of the training is completed when compared to the state-of-the-art offloading approach in FL. FedFly has negligible overhead of up to two seconds and does not compromise accuracy. Finally, we highlight a number of open research issues for further investigation. FedFly can be downloaded from https://github.com/qub-bless-on/FedFly.

[1]  Di Wu,et al.  FedAdapt: Adaptive Offloading for IoT Devices in Federated Learning , 2021, IEEE Internet of Things Journal.

[2]  Urmish Thakker,et al.  A Survey on Federated Learning for Resource-Constrained IoT Devices , 2021, IEEE Internet of Things Journal.

[3]  Seyit Camtepe,et al.  SplitFed: When Federated Learning Meets Split Learning , 2020, AAAI.

[4]  Aruna Seneviratne,et al.  Federated Learning for Internet of Things: A Comprehensive Survey , 2021, IEEE Communications Surveys & Tutorials.

[5]  Zeyi Tao,et al.  A survey of federated learning for edge computing: Research problems and solutions , 2021, High-Confidence Computing.

[6]  Parminder Singh,et al.  A Taxonomy and Survey on Container Migration Techniques in Cloud Computing , 2021 .

[7]  Blesson Varghese,et al.  Scission: Performance-driven and Context-aware Cloud-Edge Distribution of Deep Neural Networks , 2020, 2020 IEEE/ACM 13th International Conference on Utility and Cloud Computing (UCC).

[8]  Surya Nepal,et al.  End-to-End Evaluation of Federated Learning and Split Learning for Internet of Things , 2020, 2020 International Symposium on Reliable Distributed Systems (SRDS).

[9]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[10]  Aaas News,et al.  Book Reviews , 1893, Buffalo Medical and Surgical Journal.

[11]  W. Hager,et al.  and s , 2019, Shallow Water Hydraulics.

[12]  Ramesh Raskar,et al.  Split learning for health: Distributed deep learning without sharing raw patient data , 2018, ArXiv.

[13]  Xiaoming Fu,et al.  A Survey on Virtual Machine Migration: Challenges, Techniques, and Open Issues , 2018, IEEE Communications Surveys & Tutorials.

[14]  Shripad Nadgowda,et al.  Voyager: Complete Container State Migration , 2017, 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS).

[15]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[16]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[17]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[18]  Neil Genzlinger A. and Q , 2006 .