Communication-Efficient Online Federated Learning Strategies for Kernel Regression

This article presents communication-efficient approaches to federated learning (FL) for resource-constrained devices with access to streaming data. In particular, we first propose a partial-sharing-based framework for online federated learning (PSO-Fed), wherein clients update local models from a stream of data and exchange tiny fractions of the model with the server, reducing the communication overhead. In contrast to classical FL approaches, the proposed strategy provides clients who are not part of a global iteration with the freedom to update local models whenever new data arrives. Furthermore, by devising a client-side innovation check, we also propose an event-triggered PSO-Fed (ETPSO-Fed) that further reduces the computational burden of clients while enhancing communication efficiency. We implement the above-mentioned frameworks in the context of kernel regression, where clients perform local learning employing random Fourier features (RFFs)-based kernel least mean squares. In addition, we examine the mean and mean-square convergence of the proposed PSO-Fed. Finally, we conduct experiments to determine the efficacy of the proposed frameworks. Our results show that PSO-Fed and ETPSO-Fed can compete with Online-Fed while requiring significantly less communication overhead. Simulations demonstrate an 80% reduction in PSO-Fed and an 84.5% reduction in ETPSO-Fed communication overhead compared to Online-Fed. Notably, the proposed PSO-Fed strategies show good resilience against model-poisoning attacks without involving additional mechanisms.

[1]  Vinay Chakravarthi Gogineni,et al.  Algorithm and Architecture Design of Random Fourier Features-Based Kernel Adaptive Filters , 2023, IEEE Transactions on Circuits and Systems I: Regular Papers.

[2]  Vinay Chakravarthi Gogineni,et al.  Graph Kernel Recursive Least-Squares Algorithms , 2021, 2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC).

[3]  A. Kuh Real Time Kernel Learning for Sensor Networks using Principles of Federated Learning , 2021, 2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC).

[4]  Tuo Zhang,et al.  Federated Learning for Internet of Things: Applications, Challenges, and Opportunities , 2021, ArXiv.

[5]  Xuefei Yin,et al.  A Comprehensive Survey of Privacy-preserving Federated Learning , 2021, ACM Comput. Surv..

[6]  Xueyu Wu,et al.  FedSCR: Structure-Based Communication Reduction for Federated Learning , 2021, IEEE Transactions on Parallel and Distributed Systems.

[7]  Mohsen Guizani,et al.  A Survey on Federated Learning: The Journey From Centralized to Distributed On-Site Learning and Beyond , 2021, IEEE Internet of Things Journal.

[8]  Vinay Chakravarthi Gogineni,et al.  Graph Diffusion Kernel LMS using Random Fourier Features , 2020, 2020 54th Asilomar Conference on Signals, Systems, and Computers.

[9]  Ali H. Sayed,et al.  Optimal Importance Sampling for Federated Learning , 2020, ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[10]  Huzefa Rangwala,et al.  FedAT: A Communication-Efficient Federated Learning Method with Asynchronous Tiers under Non-IID Data , 2020, ArXiv.

[11]  Mung Chiang,et al.  Fast-Convergent Federated Learning , 2020, IEEE Journal on Selected Areas in Communications.

[12]  A. Salman Avestimehr,et al.  Byzantine-Resilient Secure Federated Learning , 2020, IEEE Journal on Selected Areas in Communications.

[13]  Volkan Cevher,et al.  Machine Learning From Distributed, Streaming Data [From the Guest Editors] , 2020, IEEE Signal Process. Mag..

[14]  H. Dai,et al.  Stochastic-Sign SGD for Federated Learning with Theoretical Guarantees , 2020, ArXiv.

[15]  G. Giannakis,et al.  Federated Variance-Reduced Stochastic Gradient Descent With Robustness to Byzantine Attacks , 2019, IEEE Transactions on Signal Processing.

[16]  Monica Nicoli,et al.  Federated Learning With Cooperating Devices: A Consensus Approach for Massive IoT Networks , 2019, IEEE Internet of Things Journal.

[17]  Martin Slawski,et al.  Asynchronous Online Federated Learning for Edge Devices with Non-IID Data , 2019, 2020 IEEE International Conference on Big Data (Big Data).

[18]  Wei Yang Bryan Lim,et al.  Federated Learning in Mobile Edge Networks: A Comprehensive Survey , 2019, IEEE Communications Surveys & Tutorials.

[19]  H. Poor,et al.  A Joint Learning and Communications Framework for Federated Learning Over Wireless Networks , 2019, IEEE Transactions on Wireless Communications.

[20]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[21]  H. Vincent Poor,et al.  Scheduling Policies for Federated Learning in Wireless Networks , 2019, IEEE Transactions on Communications.

[22]  Solmaz Niknam,et al.  Federated Learning for Wireless Communications: Motivation, Opportunities, and Challenges , 2019, IEEE Communications Magazine.

[23]  Wei Wang,et al.  CMFL: Mitigating Communication Overhead for Federated Learning , 2019, 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS).

[24]  Vinay Chakravarthi Gogineni,et al.  Partial Diffusion Affine Projection Algorithm Over Clustered Multitask Networks , 2019, 2019 IEEE International Symposium on Circuits and Systems (ISCAS).

[25]  Klaus-Robert Müller,et al.  Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[26]  Qiang Yang,et al.  Federated Machine Learning , 2019, ACM Trans. Intell. Syst. Technol..

[27]  Anit Kumar Sahu,et al.  Federated Optimization in Heterogeneous Networks , 2018, MLSys.

[28]  Mehdi Bennis,et al.  Wireless Network Intelligence at the Edge , 2018, Proceedings of the IEEE.

[29]  Mehdi Bennis,et al.  Communication-Efficient On-Device Machine Learning: Federated Distillation and Augmentation under Non-IID Private Data , 2018, ArXiv.

[30]  Walid Saad,et al.  Distributed Federated Learning for Ultra-Reliable Low-Latency Vehicular Communications , 2018, IEEE Transactions on Communications.

[31]  Takayuki Nishio,et al.  Client Selection for Federated Learning with Heterogeneous Resources in Mobile Edge , 2018, ICC 2019 - 2019 IEEE International Conference on Communications (ICC).

[32]  Sarvar Patel,et al.  Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..

[33]  Cédric Richard,et al.  Decentralized Online Learning With Kernels , 2017, IEEE Transactions on Signal Processing.

[34]  Yair Be'ery,et al.  Decentralized estimation of regression coefficients in sensor networks , 2017, Digit. Signal Process..

[35]  Ameet Talwalkar,et al.  Federated Multi-Task Learning , 2017, NIPS.

[36]  Sergios Theodoridis,et al.  Online Distributed Learning Over Networks in RKH Spaces Using Random Fourier Features , 2017, IEEE Transactions on Signal Processing.

[37]  Peter Richtárik,et al.  Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.

[38]  Peter Richtárik,et al.  Federated Optimization: Distributed Machine Learning for On-Device Intelligence , 2016, ArXiv.

[39]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[40]  E. Saff,et al.  Fundamentals of Matrix Analysis with Applications , 2015 .

[41]  Sergios Theodoridis,et al.  Machine Learning: A Bayesian and Optimization Perspective , 2015 .

[42]  Yih-Fang Huang,et al.  Adaptive Distributed Estimation Based on Recursive Least-Squares and Partial Diffusion , 2014, IEEE Transactions on Signal Processing.

[43]  G. Moustakides,et al.  Sequential and decentralized estimation of linear-regression parameters in wireless sensor networks , 2013, IEEE Transactions on Aerospace and Electronic Systems.

[44]  Ali H. Sayed,et al.  Diffusion Adaptation Strategies for Distributed Optimization and Learning Over Networks , 2011, IEEE Transactions on Signal Processing.

[45]  S. Haykin,et al.  Kernel Least‐Mean‐Square Algorithm , 2010 .

[46]  Weifeng Liu,et al.  Kernel Adaptive Filtering: A Comprehensive Introduction , 2010 .

[47]  Paul Honeine,et al.  Online Prediction of Time Series Data With Kernels , 2009, IEEE Transactions on Signal Processing.

[48]  Benjamin Recht,et al.  Random Features for Large-Scale Kernel Machines , 2007, NIPS.

[49]  Leonidas J. Guibas,et al.  Collaborative signal and information processing: an information-directed approach , 2003 .

[50]  Shirish Nagaraj,et al.  Set-membership filtering and a set-membership normalized LMS algorithm with an adaptive step size , 1998, IEEE Signal Processing Letters.

[51]  H. Neudecker,et al.  Block Kronecker products and the vecb operator , 1991 .

[52]  Vinay Chakravarthi Gogineni,et al.  Kernel Regression Over Graphs Using Random Fourier Features , 2022, IEEE Transactions on Signal Processing.

[53]  Hoi-To Wai,et al.  Decentralized Learning for Overparameterized Problems: A Multi-Agent Kernel Approximation Approach , 2022, ICLR.

[54]  Rui Li,et al.  Online Federated Multitask Learning , 2020 .

[55]  Yih-Fang Huang,et al.  Nonlinear Adaptive Filtering With Kernel Set-Membership Approach , 2020, IEEE Transactions on Signal Processing.

[56]  Yih-Fang Huang,et al.  Distributed Least Mean-Square Estimation With Partial Diffusion , 2014, IEEE Transactions on Signal Processing.

[57]  Yuhao Zhou,et al.  Communication-Efficient Federated Learning With Compensated Overlap-FedAvg , 2022 .