CHEESE: Distributed Clustering-Based Hybrid Federated Split Learning Over Edge Networks

Implementing either Federated learning (FL) or split learning (SL) over clients with limited computation/communication resources faces challenges on achieving delay-efficient model training. To overcome such challenges, we investigate a novel distributed Clustering-based Hybrid fEdErated Split lEarning (CHEESE) framework, consolidating distributed resources among clients by device-to-device (D2D) communications, working in an intra-serial inter-parallel manner. In CHEESE, each learning client can form a cluster with its neighboring helping clients via D2D communications to train an FL model collaboratively. Inside each cluster, the model is split into multiple segments via a model splitting and allocation (MSA) strategy, while each cluster member trains one segment. After completing intra-cluster training, a transmission client (TC) is determined from each cluster to upload a complete model to the base station for global model aggregation under allocated bandwidth. Accordingly, an overall training delay cost minimization problem is formulated, involving the following subproblems: client clustering, MSA, TC selection, and bandwidth allocation. Due to its NP-Hardness, the problem is decoupled and solved iteratively. The client clustering problem is first transformed into a distributed clustering game based on potential game theory, where each cluster further investigates the remaining three subproblems to evaluate the utility of each clustering strategy. Specifically, a heuristic algorithm is proposed to solve the MSA problem under a given clustering strategy, while a greedy-based convex optimization approach is introduced to solve the joint TC selection and bandwidth allocation problem. Extensive experiments on practical models and datasets demonstrate that CHEESE can significantly reduce training delay costs.

[1]  Long Shi,et al.  Low-Latency Federated Learning With DNN Partition in Distributed Industrial IoT Networks , 2022, IEEE Journal on Selected Areas in Communications.

[2]  Xiaojiang Du,et al.  Hybrid Architectures for Distributed Machine Learning in Heterogeneous Wireless Networks , 2022, ArXiv.

[3]  C. Chakrabarti,et al.  ResSFL: A Resistance Transfer Framework for Defending Model Inversion Attack in Split Federated Learning , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Cong Wang,et al.  Economical Behavior Modeling and Analyses for Data Collection in Edge Internet of Things Networks , 2022, ACM Trans. Sens. Networks.

[5]  W. Zhuang,et al.  Split Learning Over Wireless Networks: Parallel Design and Resource Management , 2022, IEEE Journal on Selected Areas in Communications.

[6]  Jer Shyuan Ng,et al.  Decentralized Edge Intelligence: A Dynamic Resource Allocation Framework for Hierarchical Federated Learning , 2022, IEEE Transactions on Parallel and Distributed Systems.

[7]  Lichao Sun,et al.  FedBERT: When Federated Learning Meets Pre-training , 2022, ACM Trans. Intell. Syst. Technol..

[8]  Soumaya Cherkaoui,et al.  Clustered Vehicular Federated Learning: Process and Optimization , 2022, IEEE Transactions on Intelligent Transportation Systems.

[9]  Zhu Han,et al.  Distributed Clustering-Based Cooperative Vehicular Edge Computing for Real-Time Offloading Requests , 2022, IEEE Transactions on Vehicular Technology.

[10]  Ibrahim Matta,et al.  Federated or Split? A Performance and Privacy Analysis of Hybrid Split and Federated Learning Architectures , 2021, 2021 IEEE 14th International Conference on Cloud Computing (CLOUD).

[11]  Liang Ao,et al.  Joint Task Offloading, D2D Pairing, and Resource Allocation in Device-Enhanced MEC: A Potential Game Approach , 2021, IEEE Internet of Things Journal.

[12]  Xueyu Wu,et al.  FedSCR: Structure-Based Communication Reduction for Federated Learning , 2021, IEEE Transactions on Parallel and Distributed Systems.

[13]  Christopher G. Brinton,et al.  Semi-Decentralized Federated Learning With Cooperative D2D Local Model Aggregations , 2021, IEEE Journal on Selected Areas in Communications.

[14]  Virginia Smith,et al.  Label Leakage and Protection in Two-party Split Learning , 2021, ICLR.

[15]  Zhu Han,et al.  Matching-Theory-Based Low-Latency Scheme for Multitask Federated Learning in MEC Networks , 2021, IEEE Internet of Things Journal.

[16]  Choong Seon Hong,et al.  Toward Multiple Federated Learning Services Resource Sharing in Mobile Edge Networks , 2020, IEEE Transactions on Mobile Computing.

[17]  Seyit A. Camtepe,et al.  Advancements of federated learning towards privacy preservation: from federated learning to split learning , 2020, Federated Learning Systems.

[18]  Xin Liu,et al.  Adaptive Federated Dropout: Improving Communication Efficiency and Generalization for Federated Learning , 2020, IEEE INFOCOM 2021 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[19]  Shehroz S. Khan,et al.  FedSL: Federated Split Learning on Distributed Sequential Data in Recurrent Neural Networks , 2020, Multim. Tools Appl..

[20]  Palanisamy Vigneshwaran,et al.  Cluster Based Multi Agent System for Breadth First Search , 2020, 2020 20th International Conference on Advances in ICT for Emerging Regions (ICTer).

[21]  Surya Nepal,et al.  End-to-End Evaluation of Federated Learning and Split Learning for Internet of Things , 2020, 2020 International Symposium on Reliable Distributed Systems (SRDS).

[22]  Huaiyu Dai,et al.  A Truthful Auction for Graph Job Allocation in Vehicular Cloud-Assisted Networks , 2020, IEEE Transactions on Mobile Computing.

[23]  Joohyung Jeon,et al.  Privacy-Sensitive Parallel Split Learning , 2020, 2020 International Conference on Information Networking (ICOIN).

[24]  Shiqiang Wang,et al.  Model Pruning Enables Efficient Federated Learning on Edge Devices , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[25]  Ramesh Raskar,et al.  Detailed comparison of communication efficiency of split learning and federated learning , 2019, ArXiv.

[26]  S. H. Song,et al.  Client-Edge-Cloud Hierarchical Federated Learning , 2019, ICC 2020 - 2020 IEEE International Conference on Communications (ICC).

[27]  Zbigniew Dziong,et al.  Stable femtocells cluster formation and resource allocation based on cooperative game theory , 2019, Comput. Commun..

[28]  Kin K. Leung,et al.  Adaptive Federated Learning in Resource Constrained Edge Computing Systems , 2018, IEEE Journal on Selected Areas in Communications.

[29]  Romain Dolbeau,et al.  Theoretical peak FLOPS per instruction set: a tutorial , 2017, The Journal of Supercomputing.

[30]  Yong Huat Chew,et al.  Potential Game Theory: Applications in Radio Resource Allocation , 2016 .

[31]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[32]  Shudong Jin,et al.  Prediction or Not? An Energy-Efficient Framework for Clustering-Based Data Collection in Wireless Sensor Networks , 2011, IEEE Transactions on Parallel and Distributed Systems.

[33]  Michael Barbehenn,et al.  A Note on the Complexity of Dijkstra's Algorithm for Graphs with Weighted Vertices , 1998, IEEE Trans. Computers.

[34]  Jie Wu,et al.  LightFed: An Efficient and Secure Federated Edge Learning System on Model Splitting , 2021, IEEE Transactions on Parallel and Distributed Systems.