Attack Allocation on Remote State Estimation in Multi-Systems: Structural Results and Asymptotic Solution

This paper considers optimal attack attention allocation on remote state estimation in multi-systems. Suppose there are $\mathtt{M}$ independent systems, each of which has a remote sensor monitoring the system and sending its local estimates to a fusion center over a packet-dropping channel. An attacker may generate noises to exacerbate the communication channels between sensors and the fusion center. Due to capacity limitation, at each time the attacker can exacerbate at most $\mathtt{N}$ of the $\mathtt{M}$ channels. The goal of the attacker side is to seek an optimal policy maximizing the estimation error at the fusion center. The problem is formulated as a Markov decision process (MDP) problem, and the existence of an optimal deterministic and stationary policy is proved. We further show that the optimal policy has a threshold structure, by which the computational complexity is reduced significantly. Based on the threshold structure, a myopic policy is proposed for homogeneous models and its optimality is established. To overcome the curse of dimensionality of MDP algorithms for general heterogeneous models, we further provide an asymptotically (as $\mathtt{M}$ and $\mathtt{N}$ go to infinity) optimal solution, which is easy to compute and implement. Numerical examples are given to illustrate the main results.

[1]  Rafal Rohozinski,et al.  Stuxnet and the Future of Cyber War , 2011 .

[2]  Karl Henrik Johansson,et al.  Attack models and scenarios for networked control systems , 2012, HiCoNS '12.

[3]  Roy S. Smith,et al.  A Decoupled Feedback Structure for Covertly Appropriating Networked Control Systems , 2011 .

[4]  Ling Shi,et al.  Dynamic sensor transmission power scheduling for remote state estimation , 2014, Autom..

[5]  Quanxin Zhu,et al.  Value iteration for average cost Markov decision processes inBorel spaces , 2005 .

[6]  Bruno Sinopoli,et al.  Secure control against replay attacks , 2009, 2009 47th Annual Allerton Conference on Communication, Control, and Computing (Allerton).

[7]  Karl Henrik Johansson,et al.  Secure Control Systems: A Quantitative Risk Management Approach , 2015, IEEE Control Systems.

[8]  L. Sennott Stochastic Dynamic Programming and the Control of Queueing Systems , 1998 .

[9]  Quanxin Zhu,et al.  Average optimality for Markov decision processes in borel spaces: a new condition and approach , 2006, Journal of Applied Probability.

[10]  S. Meyn,et al.  Computable Bounds for Geometric Convergence Rates of Markov Chains , 1994 .

[11]  Ling Shi,et al.  Optimal Linear Cyber-Attack on Remote State Estimation , 2017, IEEE Transactions on Control of Network Systems.

[12]  Ling Shi,et al.  Infinite-horizon sensor scheduling for estimation over lossy networks , 2012, 2012 IEEE 51st IEEE Conference on Decision and Control (CDC).

[13]  Tamer Basar,et al.  Optimal control in the presence of an intelligent jammer with limited actions , 2010, 49th IEEE Conference on Decision and Control (CDC).

[14]  Richard L. Tweedie,et al.  Markov Chains and Stochastic Stability , 1993, Communications and Control Engineering Series.

[15]  Wojciech A. Trybulec Partially Ordered Sets , 1990 .

[16]  Ling Shi,et al.  Jamming Attacks on Remote State Estimation in Cyber-Physical Systems: A Game-Theoretic Approach , 2015, IEEE Transactions on Automatic Control.

[17]  Ling Shi,et al.  Optimal Denial-of-Service Attack Scheduling With Energy Constraint , 2015, IEEE Transactions on Automatic Control.

[18]  S. Shankar Sastry,et al.  Safe and Secure Networked Control Systems under Denial-of-Service Attacks , 2009, HSCC.

[19]  Francesco Bullo,et al.  Control-Theoretic Methods for Cyberphysical Security: Geometric Principles for Optimal Cross-Layer Resilient Control Systems , 2015, IEEE Control Systems.

[20]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[21]  Xiaoqiang Ren,et al.  Optimal DoS attacks on Bayesian quickest change detection , 2014, 53rd IEEE Conference on Decision and Control.

[22]  R. Weber,et al.  On an index policy for restless bandits , 1990, Journal of Applied Probability.

[23]  P. Whittle Restless Bandits: Activity Allocation in a Changing World , 1988 .

[24]  E. Byres,et al.  The Myths and Facts behind Cyber Security Risks for Industrial Control Systems , 2004 .

[25]  O. Hernández-Lerma,et al.  Further topics on discrete-time Markov control processes , 1999 .

[26]  Daniel E. Quevedo,et al.  On the optimality of threshold policies in event triggered estimation with packet drops , 2015, 2015 European Control Conference (ECC).