CodedPaddedFL and CodedSecAgg: Straggler Mitigation and Secure Aggregation in Federated Learning

We present two novel federated learning (FL) schemes that mitigate the effect of straggling devices by introducing redundancy on the devices’ data across the network. Compared to other schemes in the literature, which deal with stragglers or device dropouts by ignoring their contribution, the proposed schemes do not suffer from the client drift problem. The first scheme, CodedPaddedFL, mitigates the effect of stragglers while retaining the privacy level of conventional FL. It combines one-time padding for user data privacy with gradient codes to yield straggler resiliency. The second scheme, CodedSecAgg, provides straggler resiliency and robustness against model inversion attacks and is based on Shamir’s secret sharing. We apply CodedPaddedFL and CodedSecAgg to a classification problem. For a scenario with 120 devices, CodedPaddedFL achieves a speed-up factor of 18 for an accuracy of 95% on the MNIST dataset compared to conventional FL. Furthermore, it yields similar performance in terms of latency compared to a recently proposed scheme by Prakash et al. without the shortcoming of additional leakage of private data. CodedSecAgg outperforms the state-of-the-art secure aggregation scheme LightSecAgg by a speed-up factor of 6.6–18.7 for the MNIST dataset for an accuracy of 95%.

[1]  G. Caire,et al.  SwiftAgg+: Achieving Asymptotically Optimal Communication Loads in Secure Aggregation for Federated Learning , 2022, IEEE Journal on Selected Areas in Communications.

[2]  E. Rosnes,et al.  Straggler-Resilient Secure Aggregation for Federated Learning , 2022, 2022 30th European Signal Processing Conference (EUSIPCO).

[3]  G. Caire,et al.  SwiftAgg: Communication-Efficient and Dropout-Resistant Secure Aggregation for Federated Learning with Worst-Case Security Guarantees , 2022, 2022 IEEE International Symposium on Information Theory (ISIT).

[4]  L. V. D. Maaten,et al.  EIFFeL: Ensuring Integrity for Federated Learning , 2021, CCS.

[5]  A. Graell i Amat,et al.  Privacy-Preserving Coded Mobile Edge Computing for Low-Latency Distributed Inference , 2021, IEEE Journal on Selected Areas in Communications.

[6]  E. Rosnes,et al.  Coding for Straggler Mitigation in Federated Learning , 2021, ICC 2022 - IEEE International Conference on Communications.

[7]  Ramy E. Ali,et al.  LightSecAgg: a Lightweight and Versatile Design for Secure Aggregation in Federated Learning , 2021, MLSys.

[8]  Michael G. Rabbat,et al.  Federated Learning with Buffered Asynchronous Aggregation , 2021, AISTATS.

[9]  A. Elkordy,et al.  HeteroSAg: Secure Aggregation With Heterogeneous Quantization in Federated Learning , 2020, IEEE Transactions on Communications.

[10]  Ramy E. Ali,et al.  Secure Aggregation for Buffered Asynchronous Federated Learning , 2021, ArXiv.

[11]  Alexandre Graell i Amat,et al.  Low-Latency Distributed Inference at the Network Edge Using Rateless Codes (Invited Paper) , 2021, 2021 17th International Symposium on Wireless Communication Systems (ISWCS).

[12]  Alexandre Graell i Amat,et al.  Rateless Codes for Low-Latency Distributed Inference in Mobile Edge Computing , 2021, ArXiv.

[13]  George J. Pappas,et al.  Linear Convergence in Federated Learning: Tackling Client Heterogeneity and Sparse Gradients , 2021, NeurIPS.

[14]  Yizhou Zhao,et al.  Information Theoretic Secure Aggregation with User Dropouts , 2021, 2021 IEEE International Symposium on Information Theory (ISIT).

[15]  Nageen Himayat,et al.  Coded Computing for Low-Latency Federated Learning Over Wireless Edge Networks , 2020, IEEE Journal on Selected Areas in Communications.

[16]  A. Salman Avestimehr,et al.  Turbo-Aggregate: Breaking the Quadratic Aggregation Barrier in Secure Federated Learning , 2020, IEEE Journal on Selected Areas in Information Theory.

[17]  Stephen A. Jarvis,et al.  SAFA: A Semi-Asynchronous Protocol for Fast Federated Learning With Low Overhead , 2019, IEEE Transactions on Computers.

[18]  Cuntai Guan,et al.  A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Amir Salman Avestimehr,et al.  LightSecAgg: Rethinking Secure Aggregation in Federated Learning , 2021, ArXiv.

[20]  George J. Pappas,et al.  Achieving Linear Convergence in Federated Learning under Objective and Systems Heterogeneity , 2021, ArXiv.

[21]  Tancrède Lepoint,et al.  Secure Single-Server Aggregation with (Poly)Logarithmic Overhead , 2020, IACR Cryptol. ePrint Arch..

[22]  Kannan Ramchandran,et al.  FastSecAgg: Scalable Secure Aggregation for Privacy-Preserving Federated Learning , 2020, ArXiv.

[23]  Qinghua Liu,et al.  Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization , 2020, NeurIPS.

[24]  Jakub Konecný,et al.  On the Outsized Importance of Learning Rates in Local Update Methods , 2020, ArXiv.

[25]  Alexandre Graell i Amat,et al.  Private Edge Computing for Linear Inference Based on Secret Sharing , 2020, GLOBECOM 2020 - 2020 IEEE Global Communications Conference.

[26]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[27]  Anit Kumar Sahu,et al.  Federated Optimization in Heterogeneous Networks , 2018, MLSys.

[28]  Kan Yang,et al.  VerifyNet: Secure and Verifiable Federated Learning , 2020, IEEE Transactions on Information Forensics and Security.

[29]  Shusen Yang,et al.  Asynchronous Federated Learning with Differential Privacy for Edge Intelligence , 2019, ArXiv.

[30]  Osvaldo Simeone,et al.  On Model Coding for Distributed Inference and Transmission in Mobile Edge Computing Systems , 2019, IEEE Communications Letters.

[31]  Indranil Gupta,et al.  Asynchronous Federated Optimization , 2019, ArXiv.

[32]  Hubert Eichner,et al.  Towards Federated Learning at Scale: System Design , 2019, SysML.

[33]  Yang Song,et al.  Beyond Inferring Class Representatives: User-Level Privacy Leakage From Federated Learning , 2018, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[34]  Amir Salman Avestimehr,et al.  Lagrange Coded Computing: Optimal Design for Resiliency, Security and Privacy , 2018, AISTATS.

[35]  Albin Severinson,et al.  Block-Diagonal and LT Codes for Distributed Computing With Straggling Servers , 2017, IEEE Transactions on Communications.

[36]  Pulkit Grover,et al.  “Short-Dot”: Computing Large Linear Transforms Distributedly Using Coded Short Dot Products , 2017, IEEE Transactions on Information Theory.

[37]  Kannan Ramchandran,et al.  Speeding Up Distributed Machine Learning Using Codes , 2015, IEEE Transactions on Information Theory.

[38]  Suhas N. Diggavi,et al.  Straggler Mitigation in Distributed Optimization Through Data Encoding , 2017, NIPS.

[39]  Sarvar Patel,et al.  Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..

[40]  P. Lambin,et al.  Developing and Validating a Survival Prediction Model for NSCLC Patients Through Distributed Learning Across 3 Countries , 2017, International journal of radiation oncology, biology, physics.

[41]  Farzin Haddadpour,et al.  On the optimal recovery threshold of coded matrix multiplication , 2017, 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[42]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[43]  Alexandros G. Dimakis,et al.  Gradient Coding: Avoiding Stragglers in Distributed Learning , 2017, ICML.

[44]  Mohammad Ali Maddah-Ali,et al.  Polynomial Codes: an Optimal Design for High-Dimensional Coded Matrix Multiplication , 2017, NIPS.

[45]  Avanti Shrikumar,et al.  Learning Important Features Through Propagating Activation Differences , 2017, ICML.

[46]  Amir Salman Avestimehr,et al.  Coded computation over heterogeneous clusters , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).

[47]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[48]  Peter Richtárik,et al.  Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.

[49]  Mohammad Ali Maddah-Ali,et al.  A Unified Coding Framework for Distributed Computing with Straggling Servers , 2016, 2016 IEEE Globecom Workshops (GC Wkshps).

[50]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[51]  Somesh Jha,et al.  Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.

[52]  Octavian Catrina,et al.  Secure Computation with Fixed-Point Numbers , 2010, Financial Cryptography.

[53]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[54]  Adi Shamir,et al.  How to share a secret , 1979, CACM.

[55]  Claude E. Shannon,et al.  Communication theory of secrecy systems , 1949, Bell Syst. Tech. J..