Byzantine Fault Tolerant Distributed Linear Regression

This paper considers the problem of Byzantine fault tolerance in distributed linear regression in a multi-agent system. However, the proposed algorithms are given for a more general class of distributed optimization problems, of which distributed linear regression is a special case. The system comprises of a server and multiple agents, where each agent is holding a certain number of data points and responses that satisfy a linear relationship (could be noisy). The objective of the server is to determine this relationship, given that some of the agents in the system (up to a known number) are Byzantine faulty (aka. actively adversarial). We show that the server can achieve this objective, in a deterministic manner, by robustifying the original distributed gradient descent method using norm based filters, namely 'norm filtering' and 'norm-cap filtering', incurring an additional log-linear computation cost in each iteration. The proposed algorithms improve upon the existing methods on three levels: i) no assumptions are required on the probability distribution of data points, ii) system can be partially asynchronous, and iii) the computational overhead (in order to handle Byzantine faulty agents) is log-linear in number of agents and linear in dimension of data points. The proposed algorithms differ from each other in the assumptions made for their correctness, and the gradient filter they use.

[1]  Paulo Tabuada,et al.  Secure State Estimation for Cyber-Physical Systems Under Sensor Attacks: A Satisfiability Modulo Theory Approach , 2014, IEEE Transactions on Automatic Control.

[2]  Lili Su,et al.  Distributed Statistical Machine Learning in Adversarial Settings: Byzantine Gradient Descent , 2017, Abstracts of the 2018 ACM International Conference on Measurement and Modeling of Computer Systems.

[3]  Razvan Pascanu,et al.  Understanding the exploding gradient problem , 2012, ArXiv.

[4]  W. Rudin Principles of mathematical analysis , 1964 .

[5]  Dan Alistarh,et al.  Byzantine Stochastic Gradient Descent , 2018, NeurIPS.

[6]  Paulo Tabuada,et al.  Robustness of attack-resilient state estimators , 2014, 2014 ACM/IEEE International Conference on Cyber-Physical Systems (ICCPS).

[7]  Indranil Gupta,et al.  Generalized Byzantine-tolerant SGD , 2018, ArXiv.

[8]  Allen Y. Yang,et al.  Robust Face Recognition via Sparse Representation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  L. Eon Bottou Online Learning and Stochastic Approximations , 1998 .

[10]  Léon Bottou,et al.  On-line learning and stochastic approximations , 1999 .

[11]  Sivaraman Balakrishnan,et al.  Robust estimation via robust gradient estimation , 2018, Journal of the Royal Statistical Society: Series B (Statistical Methodology).

[12]  Nitin H. Vaidya,et al.  Fault-Tolerant Multi-Agent Optimization: Optimal Iterative Distributed Algorithms , 2016, PODC.

[13]  Insup Lee,et al.  Attack-Resilient State Estimation for Noisy Dynamical Systems , 2017, IEEE Transactions on Control of Network Systems.

[14]  Rachid Guerraoui,et al.  Asynchronous Byzantine Machine Learning ( the case of SGD ) Supplementary Material , 2022 .

[15]  Leslie Lamport,et al.  The Byzantine Generals Problem , 1982, TOPL.

[16]  Wade Trappe,et al.  Robust statistical methods for securing wireless localization in sensor networks , 2005, IPSN 2005. Fourth International Symposium on Information Processing in Sensor Networks, 2005..

[17]  John N. Tsitsiklis,et al.  Parallel and distributed computation , 1989 .

[18]  Yinyu Ye,et al.  An extension of Karmarkar's projective algorithm for convex quadratic programming , 1989, Math. Program..

[19]  Suhas N. Diggavi,et al.  Data Encoding for Byzantine-Resilient Distributed Gradient Descent , 2018, 2018 56th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[20]  Lili Su,et al.  Finite-Time Guarantees for Byzantine-Resilient Distributed State Estimation With Noisy Measurements , 2018, IEEE Transactions on Automatic Control.

[21]  Paulo Tabuada,et al.  Secure Estimation and Control for Cyber-Physical Systems Under Adversarial Attacks , 2012, IEEE Transactions on Automatic Control.

[22]  Vitaly Shmatikov,et al.  Privacy-preserving deep learning , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[23]  Nitin H. Vaidya,et al.  Multi-agent optimization in the presence of Byzantine adversaries: Fundamental limits , 2016, 2016 American Control Conference (ACC).

[24]  Joachim M. Buhmann,et al.  Fast and Robust Least Squares Estimation in Corrupted Linear Models , 2014, NIPS.

[25]  Soummya Kar,et al.  Resilient Distributed Estimation Through Adversary Detection , 2017, IEEE Transactions on Signal Processing.

[26]  Kamyar Azizzadenesheli,et al.  signSGD with Majority Vote is Communication Efficient and Fault Tolerant , 2018, ICLR.

[27]  Qing Ling,et al.  Robust decentralized dynamic optimization at presence of malfunctioning agents , 2018, Signal Process..

[28]  B. Gharesifard,et al.  Distributed Optimization Under Adversarial Nodes , 2016, IEEE Transactions on Automatic Control.

[29]  Xiaoqiang Ren,et al.  Secure State Estimation With Byzantine Sensors: A Probabilistic Approach , 2019, IEEE Transactions on Automatic Control.

[30]  Nitin H. Vaidya,et al.  Robust Multi-agent Optimization: Coping with Byzantine Agents with Input Redundancy , 2016, SSS.

[31]  Waheed Uz Zaman Bajwa,et al.  ByRDiE: Byzantine-Resilient Distributed Coordinate Descent for Decentralized Learning , 2017, IEEE Transactions on Signal and Information Processing over Networks.

[32]  Prateek Jain,et al.  Robust Regression via Hard Thresholding , 2015, NIPS.

[33]  Jerry Li,et al.  Sever: A Robust Meta-Algorithm for Stochastic Optimization , 2018, ICML.

[34]  Kannan Ramchandran,et al.  Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates , 2018, ICML.

[35]  Shie Mannor,et al.  Robust Sparse Regression under Adversarial Corruption , 2013, ICML.

[36]  Kamyar Azizzadenesheli,et al.  signSGD with Majority Vote is Communication Efficient And Byzantine Fault Tolerant , 2018, ArXiv.

[37]  Lifeng Lai,et al.  Distributed Gradient Descent Algorithm Robust to an Arbitrary Number of Byzantine Attackers , 2019, IEEE Transactions on Signal Processing.

[38]  Rachid Guerraoui,et al.  Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent , 2017, NIPS.

[39]  Farzam Fanitabasi,et al.  A Review of Adversarial Behaviour in Distributed Multi-Agent Optimisation , 2018, 2018 IEEE/ACM International Conference on Utility and Cloud Computing Companion (UCC Companion).

[40]  Jiannong Cao,et al.  Secure localization and location verification in wireless sensor networks: a survey , 2010, The Journal of Supercomputing.

[41]  João Pedro Hespanha,et al.  Observability of linear systems under adversarial attacks , 2015, 2015 American Control Conference (ACC).

[42]  Shreyas Sundaram,et al.  Byzantine-resilient distributed observers for LTI systems , 2018, Autom..