FlexiFed: Personalized Federated Learning for Edge Clients with Heterogeneous Model Architectures

Mobile and Web-of-Things (WoT) devices at the network edge account for more than half of the world’s web traffic, making a great data source for various machine learning (ML) applications, particularly federated learning (FL) which offers a promising solution to privacy-preserving ML feeding on these data. FL allows edge mobile and WoT devices to train a shared global ML model under the orchestration of a central parameter server. In the real world, due to resource heterogeneity, these edge devices often train different versions of models (e.g., VGG-16 and VGG-19) or different ML models (e.g., VGG and ResNet) for the same ML task (e.g., computer vision and speech recognition). Existing FL schemes have assumed that participating edge devices share a common model architecture, and thus cannot facilitate FL across edge devices with heterogeneous ML model architectures. We explored this architecture heterogeneity challenge and found that FL can and should accommodate these edge devices to improve model accuracy and accelerate model training. This paper presents our findings and FlexiFed, a novel scheme for FL across edge devices with heterogeneous model architectures, and three model aggregation strategies for accommodating architecture heterogeneity under FlexiFed. Experiments with four widely-used ML models on four public datasets demonstrate 1) the usefulness of FlexiFed; and 2) that compared with the state-of-the-art FL scheme, FlexiFed improves model accuracy by 2.6%-9.7% and accelerates model convergence by 1.24 × -4.04 ×.

[1]  Siva Theja Maguluri,et al.  Federated Reinforcement Learning: Linear Speedup Under Markovian Sampling , 2022, ICML.

[2]  Yunfeng Shao,et al.  Personalized Federated Learning via Variational Bayesian Inference , 2022, ICML.

[3]  Xinmei Tian,et al.  DisPFL: Towards Communication-Efficient Personalized Federated Learning via Decentralized Sparse Training , 2022, ICML.

[4]  R. Raskar,et al.  LocFedMix-SL: Localize, Federate, and Mix for Improved Scalability, Convergence, and Latency in Split Learning , 2022, WWW.

[5]  Michael G. Rabbat,et al.  Federated Learning with Partial Model Personalization , 2022, ICML.

[6]  Liang Ding,et al.  Fine-tuning Global Model via Data-Free Knowledge Distillation for Non-IID Federated Learning , 2022, Computer Vision and Pattern Recognition.

[7]  Yuanqing Lin,et al.  Improving Federated Learning Face Recognition via Privacy-Agnostic Clusters , 2022, ICLR.

[8]  Mohammad Abdizadeh,et al.  Federated Learning With Taskonomy for Non-IID Data , 2021, IEEE Transactions on Neural Networks and Learning Systems.

[9]  K. Ramchandran,et al.  An Efficient Framework for Clustered Federated Learning , 2020, IEEE Transactions on Information Theory.

[10]  Seyit Camtepe,et al.  SplitFed: When Federated Learning Meets Split Learning , 2020, AAAI.

[11]  Xue Yang,et al.  An Accuracy-Lossless Perturbation Method for Defending Privacy Attacks in Federated Learning , 2020, WWW.

[12]  J. Bilmes,et al.  Diverse Client Selection for Federated Learning via Submodular Maximization , 2022, ICLR.

[13]  Junchi Yan,et al.  Deep Neural Network Fusion via Graph Matching with Applications to Model Ensemble and Federated Learning , 2022, ICML.

[14]  H. Li,et al.  Hermes: an efficient federated learning framework for heterogeneous mobile clients , 2021, MobiCom.

[15]  Jiayu Zhou,et al.  Data-Free Knowledge Distillation for Heterogeneous Federated Learning , 2021, ICML.

[16]  Junhao Wang,et al.  Sample-level Data Selection for Federated Learning , 2021, IEEE INFOCOM 2021 - IEEE Conference on Computer Communications.

[17]  Fabio Galasso,et al.  Cluster-driven Graph Federated Learning over Multiple Domains , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[18]  Enhong Chen,et al.  Hierarchical Personalized Federated Learning for User Modeling , 2021, WWW.

[19]  Chenglin Li,et al.  Meta-HAR: Federated Representation Learning for Human Activity Recognition , 2021, WWW.

[20]  Yao Guo,et al.  PFA: Privacy-preserving Federated Adaptation for Effective Model Personalization , 2021, WWW.

[21]  S. Shakkottai,et al.  Exploiting Shared Representations for Personalized Federated Learning , 2021, ICML.

[22]  P. Kairouz,et al.  The Distributed Discrete Gaussian Mechanism for Federated Learning with Secure Aggregation , 2021, ICML.

[23]  Lawrence Carin,et al.  FLOP: Federated Learning on Medical Datasets using Partial Networks , 2021, KDD.

[24]  Martin Jaggi,et al.  Consensus Control for Decentralized Deep Learning , 2021, ICML.

[25]  Thomas F. La Porta,et al.  Augur: Modeling the Resource Requirements of ConvNets on Mobile Devices , 2021, IEEE Transactions on Mobile Computing.

[26]  Han Hu,et al.  Robustness of on-Device Models: Adversarial Attack to Deep Learning Models on Android Apps , 2021, 2021 IEEE/ACM 43rd International Conference on Software Engineering: Software Engineering in Practice (ICSE-SEIP).

[27]  Leandros Tassiulas,et al.  Cost-Effective Federated Learning Design , 2020, IEEE INFOCOM 2021 - IEEE Conference on Computer Communications.

[28]  M. Chowdhury,et al.  Oort: Efficient Federated Learning via Guided Participant Selection , 2020, OSDI.

[29]  Hui Xiong,et al.  A Comprehensive Survey on Transfer Learning , 2019, Proceedings of the IEEE.

[30]  Wojciech Samek,et al.  Clustered Federated Learning: Model-Agnostic Distributed Multitask Optimization Under Privacy Constraints , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[31]  Yi Yang,et al.  PartialFed: Cross-Domain Personalized Federated Learning via Partial Initialization , 2021, NeurIPS.

[32]  Shiva Prasad Kasiviswanathan,et al.  Federated Learning under Arbitrary Communication Patterns , 2021, ICML.

[33]  Filip Hanzely,et al.  Lower Bounds and Optimal Algorithms for Personalized Federated Learning , 2020, NeurIPS.

[34]  Murali Annavaram,et al.  Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge , 2020, NeurIPS.

[35]  Rui Song,et al.  Causal Effect Estimation and Optimal Dose Suggestions in Mobile Health , 2020, ICML.

[36]  Zhifei Zhang,et al.  Analyzing User-Level Privacy Attack Against Federated Learning , 2020, IEEE Journal on Selected Areas in Communications.

[37]  Nguyen H. Tran,et al.  Personalized Federated Learning with Moreau Envelopes , 2020, NeurIPS.

[38]  X. Chu,et al.  FMore: An Incentive Scheme of Multi-dimensional Auction for Federated Learning in MEC , 2020, 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS).

[39]  Aryan Mokhtari,et al.  Personalized Federated Learning: A Meta-Learning Approach , 2020, ArXiv.

[40]  Yasaman Khazaeni,et al.  Federated Learning with Matched Averaging , 2020, ICLR.

[41]  Ruslan Salakhutdinov,et al.  Think Locally, Act Globally: Federated Learning with Local and Global Representations , 2020, ArXiv.

[42]  Jinyuan Jia,et al.  Local Model Poisoning Attacks to Byzantine-Robust Federated Learning , 2019, USENIX Security Symposium.

[43]  H. Vincent Poor,et al.  Federated Learning With Differential Privacy: Algorithms and Performance Analysis , 2019, IEEE Transactions on Information Forensics and Security.

[44]  Sashank J. Reddi,et al.  SCAFFOLD: Stochastic Controlled Averaging for Federated Learning , 2019, ICML.

[45]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[46]  Tian Li,et al.  Fair Resource Allocation in Federated Learning , 2019, ICLR.

[47]  S. H. Song,et al.  Client-Edge-Cloud Hierarchical Federated Learning , 2019, ICC 2020 - 2020 IEEE International Conference on Communications (ICC).

[48]  Anit Kumar Sahu,et al.  Federated Optimization in Heterogeneous Networks , 2018, MLSys.

[49]  Sunav Choudhary,et al.  Federated Learning with Personalization Layers , 2019, ArXiv.

[50]  Jakub Konecný,et al.  Improving Federated Learning Personalization via Model Agnostic Meta Learning , 2019, ArXiv.

[51]  Mona Attariyan,et al.  Parameter-Efficient Transfer Learning for NLP , 2019, ICML.

[52]  Mehdi Bennis,et al.  Wireless Network Intelligence at the Edge , 2018, Proceedings of the IEEE.

[53]  Quoc V. Le,et al.  GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism , 2018, ArXiv.

[54]  Amos J. Storkey,et al.  School of Informatics, University of Edinburgh , 2022 .

[55]  Yue Zhao,et al.  Federated Learning with Non-IID Data , 2018, ArXiv.

[56]  Charles X. Ling,et al.  Pelee: A Real-Time Object Detection System on Mobile Devices , 2018, NeurIPS.

[57]  Pete Warden,et al.  Speech Commands: A Dataset for Limited-Vocabulary Speech Recognition , 2018, ArXiv.

[58]  Wonyong Sung,et al.  Fully Neural Network Based Speech Recognition on Mobile and Embedded Devices , 2018, NeurIPS.

[59]  Agustí Verde Parera,et al.  General data protection regulation , 2018 .

[60]  Ameet Talwalkar,et al.  Federated Multi-Task Learning , 2017, NIPS.

[61]  Yann LeCun,et al.  Very Deep Convolutional Networks for Text Classification , 2016, EACL.

[62]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[63]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[64]  Xiang Zhang,et al.  Character-level Convolutional Networks for Text Classification , 2015, NIPS.

[65]  Michael I. Jordan,et al.  Learning Transferable Features with Deep Adaptation Networks , 2015, ICML.

[66]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[67]  Yoshua Bengio,et al.  How transferable are features in deep neural networks? , 2014, NIPS.