CodedPaddedFL and CodedSecAgg: Straggler Mitigation and Secure Aggregation in Federated Learning
暂无分享,去创建一个
[1] G. Caire,et al. SwiftAgg+: Achieving Asymptotically Optimal Communication Loads in Secure Aggregation for Federated Learning , 2022, IEEE Journal on Selected Areas in Communications.
[2] E. Rosnes,et al. Straggler-Resilient Secure Aggregation for Federated Learning , 2022, 2022 30th European Signal Processing Conference (EUSIPCO).
[3] G. Caire,et al. SwiftAgg: Communication-Efficient and Dropout-Resistant Secure Aggregation for Federated Learning with Worst-Case Security Guarantees , 2022, 2022 IEEE International Symposium on Information Theory (ISIT).
[4] L. V. D. Maaten,et al. EIFFeL: Ensuring Integrity for Federated Learning , 2021, CCS.
[5] A. Graell i Amat,et al. Privacy-Preserving Coded Mobile Edge Computing for Low-Latency Distributed Inference , 2021, IEEE Journal on Selected Areas in Communications.
[6] E. Rosnes,et al. Coding for Straggler Mitigation in Federated Learning , 2021, ICC 2022 - IEEE International Conference on Communications.
[7] Ramy E. Ali,et al. LightSecAgg: a Lightweight and Versatile Design for Secure Aggregation in Federated Learning , 2021, MLSys.
[8] Michael G. Rabbat,et al. Federated Learning with Buffered Asynchronous Aggregation , 2021, AISTATS.
[9] A. Elkordy,et al. HeteroSAg: Secure Aggregation With Heterogeneous Quantization in Federated Learning , 2020, IEEE Transactions on Communications.
[10] Ramy E. Ali,et al. Secure Aggregation for Buffered Asynchronous Federated Learning , 2021, ArXiv.
[11] Alexandre Graell i Amat,et al. Low-Latency Distributed Inference at the Network Edge Using Rateless Codes (Invited Paper) , 2021, 2021 17th International Symposium on Wireless Communication Systems (ISWCS).
[12] Alexandre Graell i Amat,et al. Rateless Codes for Low-Latency Distributed Inference in Mobile Edge Computing , 2021, ArXiv.
[13] George J. Pappas,et al. Linear Convergence in Federated Learning: Tackling Client Heterogeneity and Sparse Gradients , 2021, NeurIPS.
[14] Yizhou Zhao,et al. Information Theoretic Secure Aggregation with User Dropouts , 2021, 2021 IEEE International Symposium on Information Theory (ISIT).
[15] Nageen Himayat,et al. Coded Computing for Low-Latency Federated Learning Over Wireless Edge Networks , 2020, IEEE Journal on Selected Areas in Communications.
[16] A. Salman Avestimehr,et al. Turbo-Aggregate: Breaking the Quadratic Aggregation Barrier in Secure Federated Learning , 2020, IEEE Journal on Selected Areas in Information Theory.
[17] Stephen A. Jarvis,et al. SAFA: A Semi-Asynchronous Protocol for Fast Federated Learning With Low Overhead , 2019, IEEE Transactions on Computers.
[18] Cuntai Guan,et al. A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.
[19] Amir Salman Avestimehr,et al. LightSecAgg: Rethinking Secure Aggregation in Federated Learning , 2021, ArXiv.
[20] George J. Pappas,et al. Achieving Linear Convergence in Federated Learning under Objective and Systems Heterogeneity , 2021, ArXiv.
[21] Tancrède Lepoint,et al. Secure Single-Server Aggregation with (Poly)Logarithmic Overhead , 2020, IACR Cryptol. ePrint Arch..
[22] Kannan Ramchandran,et al. FastSecAgg: Scalable Secure Aggregation for Privacy-Preserving Federated Learning , 2020, ArXiv.
[23] Qinghua Liu,et al. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization , 2020, NeurIPS.
[24] Jakub Konecný,et al. On the Outsized Importance of Learning Rates in Local Update Methods , 2020, ArXiv.
[25] Alexandre Graell i Amat,et al. Private Edge Computing for Linear Inference Based on Secret Sharing , 2020, GLOBECOM 2020 - 2020 IEEE Global Communications Conference.
[26] Anit Kumar Sahu,et al. Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.
[27] Anit Kumar Sahu,et al. Federated Optimization in Heterogeneous Networks , 2018, MLSys.
[28] Kan Yang,et al. VerifyNet: Secure and Verifiable Federated Learning , 2020, IEEE Transactions on Information Forensics and Security.
[29] Shusen Yang,et al. Asynchronous Federated Learning with Differential Privacy for Edge Intelligence , 2019, ArXiv.
[30] Osvaldo Simeone,et al. On Model Coding for Distributed Inference and Transmission in Mobile Edge Computing Systems , 2019, IEEE Communications Letters.
[31] Indranil Gupta,et al. Asynchronous Federated Optimization , 2019, ArXiv.
[32] Hubert Eichner,et al. Towards Federated Learning at Scale: System Design , 2019, SysML.
[33] Yang Song,et al. Beyond Inferring Class Representatives: User-Level Privacy Leakage From Federated Learning , 2018, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.
[34] Amir Salman Avestimehr,et al. Lagrange Coded Computing: Optimal Design for Resiliency, Security and Privacy , 2018, AISTATS.
[35] Albin Severinson,et al. Block-Diagonal and LT Codes for Distributed Computing With Straggling Servers , 2017, IEEE Transactions on Communications.
[36] Pulkit Grover,et al. “Short-Dot”: Computing Large Linear Transforms Distributedly Using Coded Short Dot Products , 2017, IEEE Transactions on Information Theory.
[37] Kannan Ramchandran,et al. Speeding Up Distributed Machine Learning Using Codes , 2015, IEEE Transactions on Information Theory.
[38] Suhas N. Diggavi,et al. Straggler Mitigation in Distributed Optimization Through Data Encoding , 2017, NIPS.
[39] Sarvar Patel,et al. Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..
[40] P. Lambin,et al. Developing and Validating a Survival Prediction Model for NSCLC Patients Through Distributed Learning Across 3 Countries , 2017, International journal of radiation oncology, biology, physics.
[41] Farzin Haddadpour,et al. On the optimal recovery threshold of coded matrix multiplication , 2017, 2017 55th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[42] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[43] Alexandros G. Dimakis,et al. Gradient Coding: Avoiding Stragglers in Distributed Learning , 2017, ICML.
[44] Mohammad Ali Maddah-Ali,et al. Polynomial Codes: an Optimal Design for High-Dimensional Coded Matrix Multiplication , 2017, NIPS.
[45] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[46] Amir Salman Avestimehr,et al. Coded computation over heterogeneous clusters , 2017, 2017 IEEE International Symposium on Information Theory (ISIT).
[47] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[48] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[49] Mohammad Ali Maddah-Ali,et al. A Unified Coding Framework for Distributed Computing with Straggling Servers , 2016, 2016 IEEE Globecom Workshops (GC Wkshps).
[50] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[51] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[52] Octavian Catrina,et al. Secure Computation with Fixed-Point Numbers , 2010, Financial Cryptography.
[53] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[54] Adi Shamir,et al. How to share a secret , 1979, CACM.
[55] Claude E. Shannon,et al. Communication theory of secrecy systems , 1949, Bell Syst. Tech. J..