Cyber-Attack Recovery Strategy for Smart Grid Based on Deep Reinforcement Learning

The integration of cyber-physical system increases the vulnerabilities of critical power infrastructures. Once the malicious attackers take the substation control authorities, they can trip all the transmission lines to block the power transfer. As a consequence, asynchrony will emerge between the separated regions which had been interconnected by these transmission lines. In order to recover from the attack, a straightforward way is to reclose these transmission lines once we detect the attack. However, this may cause severe impacts on the power system, such as current inrush and power swing. Therefore, it is critical to properly choose the reclosing time to mitigate these impacts. In this paper, we propose a recovery strategy to reclose the tripped transmission lines at the optimal reclosing time. In particular, a deep reinforcement learning (RL) framework is adopted to endow the strategy with the adaptability of uncertain cyber-attack scenarios and the ability of real-time decision-making. In this framework, an environment is established to simulate the power system dynamics during the attack-recovery process and generate the training data. With these data, the deep RL based strategy can be trained to determine the optimal reclosing time. Numerical results show that the proposed strategy can minimize the cyber-attack impacts under different scenarios.

[1]  Hsiao-Dong Chiang,et al.  Theoretical foundation of the BCU method for direct stability analysis of network-reduction power system. Models with small transfer conductances , 1995 .

[2]  Thoshitha T. Gamage,et al.  Analyzing the Cyber-Physical Impact of Cyber Events on the Power Grid , 2015, IEEE Transactions on Smart Grid.

[3]  Saleh Soltan,et al.  EXPOSE the Line Failures Following a Cyber-Physical Attack on the Power Grid , 2019, IEEE Transactions on Control of Network Systems.

[4]  Zhehan Yi,et al.  Deep-Reinforcement-Learning-Based Autonomous Voltage Control for Power Grid Operations , 2020, IEEE Transactions on Power Systems.

[5]  Rafal Rohozinski,et al.  Stuxnet and the Future of Cyber War , 2011 .

[6]  G. Manimaran,et al.  Integrated Anomaly Detection for Cyber Security of the Substations , 2014, IEEE Transactions on Smart Grid.

[7]  Lingfeng Wang,et al.  Power System Reliability Analysis With Intrusion Tolerance in SCADA Systems , 2016, IEEE Transactions on Smart Grid.

[8]  Fernando L. Alvarado,et al.  Testing of Trapezoidal Integration With Damping for the Solution of Power Transient Problems , 1983, IEEE Power Engineering Review.

[9]  Zhiwei Wang,et al.  A Deep Reinforcement Learning Based Approach for Optimal Active Power Dispatch , 2019, 2019 IEEE Sustainable Power and Energy Conference (iSPEC).

[10]  Deepa Kundur,et al.  Financially Motivated FDI on SCED in Real-Time Electricity Markets: Attacks and Mitigation , 2019, IEEE Transactions on Smart Grid.

[11]  Felix F. Wu,et al.  A BCU method for direct analysis of power system transient stability , 1994 .

[12]  The Year in Review , 1990, Bio/Technology.

[13]  Francesco Bullo,et al.  Synchronization of Power Networks: Network Reduction and Effective Resistance , 2010 .

[14]  Mahesh S. Illindala,et al.  A tri-level optimization model to mitigate coordinated attacks on electric power systems in a cyber-physical environment , 2019, Applied Energy.

[15]  Yan Xu,et al.  Data-Driven Load Frequency Control for Stochastic Power Systems: A Deep Reinforcement Learning Method With Continuous Action Search , 2019, IEEE Transactions on Power Systems.

[16]  Deepa Kundur,et al.  A Coordinated Multi-Switch Attack for Cascading Failures in Smart Grid , 2014, IEEE Transactions on Smart Grid.

[17]  Chong Wang,et al.  Resilience Enhancement With Sequentially Proactive Operation Strategies , 2017, IEEE Transactions on Power Systems.

[18]  Zhehan Yi,et al.  Reinforcement-Learning-Based Optimal Control of Hybrid Energy Storage Systems in Hybrid AC–DC Microgrids , 2019, IEEE Transactions on Industrial Informatics.

[19]  Zhao Yang Dong,et al.  The 2015 Ukraine Blackout: Implications for False Data Injection Attacks , 2017, IEEE Transactions on Power Systems.

[20]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[21]  Lingkun Fu,et al.  DoS Attack Energy Management Against Remote State Estimation , 2018, IEEE Transactions on Control of Network Systems.

[22]  Chong Wang,et al.  Cyber Inference System for Substation Anomalies Against Alter-and-Hide Attacks , 2017, IEEE Transactions on Power Systems.

[23]  Vladimir Terzija,et al.  A new approach to the arcing faults detection for fast autoreclosure in transmission systems , 1995 .

[24]  Z. Chen,et al.  Computation of Optimal Reclosure Time for Transmission Lines , 2002, IEEE Power Engineering Review.

[25]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[26]  Eman Mousavinejad,et al.  A Novel Cyber Attack Detection Method in Networked Control Systems , 2018, IEEE Transactions on Cybernetics.

[27]  Athanasios V. Vasilakos,et al.  Impact Assessment of Hypothesized Cyberattacks on Interconnected Bulk Power Systems , 2018, IEEE Transactions on Smart Grid.

[28]  Shengwei Mei,et al.  Toward Efficient Cascading Outage Simulation and Probability Analysis in Power Systems , 2018, IEEE Transactions on Power Systems.

[29]  Hak-Man Kim,et al.  Double Deep $Q$ -Learning-Based Distributed Operation of Battery Energy Storage System Considering Uncertainties , 2020, IEEE Transactions on Smart Grid.

[30]  Jiming Chen,et al.  Analysis of Consensus-Based Distributed Economic Dispatch Under Stealthy Attacks , 2017, IEEE Transactions on Industrial Electronics.

[31]  Qiuwei Wu,et al.  A Numerical Approach for Hybrid Simulation of Power System Dynamics Considering Extreme Icing Events , 2018, IEEE Transactions on Smart Grid.

[32]  Kwok W. Cheung,et al.  BC Hydro's on-line transient stability assessment (TSA) model development, analysis and post-processing , 1995 .

[33]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[34]  Adeniyi A. Babalola,et al.  Real-Time Cascading Failures Prevention for Multiple Contingencies in Smart Grids Through a Multi-Agent System , 2018, IEEE Transactions on Smart Grid.

[35]  Luigi Vanfretti,et al.  An open data repository and a data processing software toolset of an equivalent Nordic grid model matched to historical electricity market data , 2017, Data in brief.

[36]  Haibo He,et al.  Model-Free Real-Time EV Charging Scheduling Based on Deep Reinforcement Learning , 2019, IEEE Transactions on Smart Grid.