Federated Continuous Learning With Broad Network Architecture

Federated learning (FL) is a machine-learning setting, where multiple clients collaboratively train a model under the coordination of a central server. The clients’ raw data are locally stored, and each client only uploads the trained weight to the server, which can mitigate the privacy risks from the centralized machine learning. However, most of the existing FL models focus on one-time learning without consideration for continuous learning. Continuous learning supports learning from streaming data continuously, so it can adapt to environmental changes and provide better real-time performance. In this article, we present a federated continuous learning scheme based on broad learning (FCL-BL) to support efficient and accurate federated continuous learning (FCL). In FCL-BL, we propose a weighted processing strategy to solve the catastrophic forgetting problem, so FCL-BL can handle continuous learning. Then, we develop a local-independent training solution to support fast and accurate training in FCL-BL. The proposed solution enables us to avoid using a time-consuming synchronous approach while addressing the inaccurate-training issue rooted in the previous asynchronous approach. Moreover, we introduce a batch-asynchronous approach and broad learning (BL) technique to guarantee the high efficiency of FCL-BL. Specifically, the batch-asynchronous approach reduces the number of client–server interaction rounds, and the BL technique supports incremental learning without retraining when learning newly produced data. Finally, theoretical analysis and experimental results further illustrate that FCL-BL is superior to the existing FL schemes in terms of efficiency and accuracy in FCL.

[1]  Jie Du,et al.  Novel Efficient RNN and LSTM-Like Architectures: Recurrent and Gated Broad Learning Systems and Their Applications for Text Classification , 2020, IEEE Transactions on Cybernetics.

[2]  C. L. Philip Chen,et al.  Broad Learning System for Control of Nonlinear Dynamic Systems , 2018, 2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[3]  Byoung-Tak Zhang,et al.  Overcoming Catastrophic Forgetting by Incremental Moment Matching , 2017, NIPS.

[4]  Boris Ginsburg,et al.  Cross-Language Transfer Learning, Continuous Learning, and Domain Adaptation for End-to-End Automatic Speech Recognition , 2020, 2005.04290.

[5]  Ning Wang,et al.  Physical Layer Key Generation in 5G Wireless Networks , 2019, IEEE Wireless Communications.

[6]  Sarvar Patel,et al.  Practical Secure Aggregation for Federated Learning on User-Held Data , 2016, ArXiv.

[7]  Ian Goodfellow,et al.  Deep Learning with Differential Privacy , 2016, CCS.

[8]  C. L. Philip Chen,et al.  Structured Manifold Broad Learning System: A Manifold Perspective for Large-Scale Chaotic Time Series Analysis and Prediction , 2019, IEEE Transactions on Knowledge and Data Engineering.

[9]  Jian Lou,et al.  An Uplink Communication-Efficient Approach to Featurewise Distributed Sparse Optimization With Differential Privacy , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[10]  Shigang Yue,et al.  Event-Driven Continuous STDP Learning With Deep Structure for Visual Pattern Recognition , 2019, IEEE Transactions on Cybernetics.

[11]  Tian Li,et al.  Fair Resource Allocation in Federated Learning , 2019, ICLR.

[12]  Michael McCloskey,et al.  Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem , 1989 .

[13]  Vitaly Shmatikov,et al.  Privacy-preserving deep learning , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[14]  Indranil Gupta,et al.  Asynchronous Federated Optimization , 2019, ArXiv.

[15]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[16]  Richard E. Turner,et al.  Partitioned Variational Inference: A unified framework encompassing federated and continual learning , 2018, ArXiv.

[17]  Sebastian Thrun,et al.  A lifelong learning perspective for mobile robot control , 1994, Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS'94).

[18]  Huzefa Rangwala,et al.  Asynchronous Online Federated Learning for Edge Devices , 2019, ArXiv.

[19]  Yue Zhao,et al.  Federated Learning with Non-IID Data , 2018, ArXiv.

[20]  C. L. Philip Chen,et al.  Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[21]  Jakub Konecný,et al.  Federated Optimization: Distributed Optimization Beyond the Datacenter , 2015, ArXiv.

[22]  Marco Scavuzzo,et al.  Asynchronous Federated Learning for Geospatial Applications , 2018, DMLE/IOTSTREAMING@PKDD/ECML.

[23]  Mohsen Guizani,et al.  Reliable Federated Learning for Mobile Networks , 2019, IEEE Wireless Communications.

[24]  Yiu-ming Cheung,et al.  Uplink Communication Efficient Differentially Private Sparse Optimization With Feature-Wise Distributed Data , 2018, AAAI.

[25]  Tie Qiu,et al.  Recurrent Broad Learning Systems for Time Series Prediction , 2020, IEEE Transactions on Cybernetics.

[26]  Hubert Eichner,et al.  Towards Federated Learning at Scale: System Design , 2019, SysML.

[27]  Jakub Konecný,et al.  Federated Learning with Autotuned Communication-Efficient Secure Aggregation , 2019, 2019 53rd Asilomar Conference on Signals, Systems, and Computers.

[28]  Nenghai Yu,et al.  Asynchronous Stochastic Gradient Descent with Delay Compensation , 2016, ICML.

[29]  Derek Hoiem,et al.  Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[30]  Shuang Feng,et al.  Fuzzy Broad Learning System: A Novel Neuro-Fuzzy Model for Regression and Classification , 2020, IEEE Transactions on Cybernetics.

[31]  Eunho Yang,et al.  Federated Continual Learning with Adaptive Parameter Communication , 2020, ArXiv.

[32]  Feng Xu,et al.  E-commerce product review sentiment classification based on a naïve Bayes continuous learning framework , 2020, Inf. Process. Manag..

[33]  Richard Nock,et al.  Advances and Open Problems in Federated Learning , 2021, Found. Trends Mach. Learn..

[34]  Razvan Pascanu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[35]  Richard Nock,et al.  Private federated learning on vertically partitioned data via entity resolution and additively homomorphic encryption , 2017, ArXiv.

[36]  Xiang Li,et al.  Communication Efficient Decentralized Training with Multiple Local Updates , 2019, ArXiv.

[37]  Ramesh Raskar,et al.  FedML: A Research Library and Benchmark for Federated Machine Learning , 2020, ArXiv.

[38]  Tassilo Klein,et al.  Differentially Private Federated Learning: A Client Level Perspective , 2017, ArXiv.

[39]  Yuanming Shi,et al.  A Quasi-Newton Method Based Vertical Federated Learning Framework for Logistic Regression , 2019, ArXiv.

[40]  C. L. Philip Chen,et al.  Discriminative graph regularized broad learning system for image recognition , 2018, Science China Information Sciences.

[41]  Walid Saad,et al.  A Joint Learning and Communications Framework for Federated Learning Over Wireless Networks , 2021, IEEE Transactions on Wireless Communications.

[42]  Zhi Ding,et al.  Federated Learning via Over-the-Air Computation , 2018, IEEE Transactions on Wireless Communications.