Safe deep reinforcement learning-based constrained optimal control scheme for active distribution networks

Abstract Reinforcement learning-based schemes are being recently applied for model-free voltage control in active distribution networks. However, existing reinforcement learning methods face challenges when it comes to continuous state and action spaces problems or problems with operation constraints. To address these limitations, this paper proposes an optimal voltage control scheme based on the safe deep reinforcement learning. In this scheme, the optimal voltage control problem is formulated as a constrained Markov decision process, in which both state and action spaces are continuous. To solve this problem efficiently, the deep deterministic policy gradient algorithm is utilized to learn the reactive power control policies, which determine the optimal control actions from the states. In contrast to existing reinforcement learning methods, deep deterministic policy gradient is naturally capable of addressing control problems with continuous state and action spaces. This is due to the utilization of deep neural networks to approximate both value function and policy. In addition, in order to handle the operation constraints in active distribution networks, a safe exploration approach is proposed to form a safety layer, which is composed directly on top the deep deterministic policy gradient actor network. This safety layer predicts the change in the constrained states and prevents the violation of active distribution networks operation constraints. Numerical simulations on modified IEEE test systems demonstrate that the proposed scheme successfully maintains all bus voltage within the allowed range, and reduces the system loss by 15% compared to the no control case.

[1]  Hongwen He,et al.  Deep reinforcement learning of energy management with continuous control strategy and traffic information for a series-parallel plug-in hybrid electric bus , 2019, Applied Energy.

[2]  Balarko Chaudhuri,et al.  Smart Loads for Voltage Control in Distribution Networks , 2017, IEEE Transactions on Smart Grid.

[3]  Gustavo Valverde,et al.  Model Predictive Control of Voltages in Active Distribution Networks , 2013, IEEE Transactions on Smart Grid.

[4]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[5]  S. Chowdhury,et al.  Microgrids and Active Distribution Networks , 2009 .

[6]  D. Karlsson,et al.  Combined Local and Remote Voltage and Reactive Power Control in the Presence of Induction Machine Distributed Generation , 2007, IEEE Transactions on Power Systems.

[7]  Geoffrey Ye Li,et al.  Deep Reinforcement Learning Based Resource Allocation for V2V Communications , 2018, IEEE Transactions on Vehicular Technology.

[8]  Peng Li,et al.  Combined decentralized and local voltage control strategy of soft open points in active distribution networks , 2019, Applied Energy.

[9]  Ali Abur,et al.  Voltage Control in Active Distribution Networks Under Uncertainty in the System Model: A Robust Optimization Approach , 2018, IEEE Transactions on Smart Grid.

[10]  Yan Xu,et al.  Data-Driven Load Frequency Control for Stochastic Power Systems: A Deep Reinforcement Learning Method With Continuous Action Search , 2019, IEEE Transactions on Power Systems.

[11]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[12]  Balasubramaniam Natarajan,et al.  State Estimation and Voltage/VAR Control in Distribution Network With Intermittent Measurements , 2014, IEEE Transactions on Smart Grid.

[13]  Chin-Teng Lin,et al.  Jointly dampening traffic oscillations and improving energy consumption with electric, connected and automated vehicles: A reinforcement learning based approach , 2020 .

[14]  R. Belmans,et al.  Generalized Steady-State VSC MTDC Model for Sequential AC/DC Power Flow Algorithms , 2012, IEEE Transactions on Power Systems.

[15]  José R. Vázquez-Canteli,et al.  Reinforcement learning for demand response: A review of algorithms and modeling techniques , 2019, Applied Energy.

[16]  Takaharu Takeshita,et al.  Line Loss Minimization in Isolated Substations and Multiple Loop Distribution Systems Using the UPFC , 2014, IEEE Transactions on Power Electronics.

[17]  Hanchen Xu,et al.  Deep Reinforcement Learning for Joint Bidding and Pricing of Load Serving Entity , 2019, IEEE Transactions on Smart Grid.

[18]  John E. Fletcher,et al.  Multi-Timescale Model Predictive Control of Battery Energy Storage System Using Conic Relaxation in Smart Distribution Grids , 2018, IEEE Transactions on Power Systems.

[19]  Felix F. Wu,et al.  Network reconfiguration in distribution systems for loss reduction and load balancing , 1989 .

[20]  Nikos D. Hatziargyriou,et al.  Distributed and Decentralized Voltage Control of Smart Distribution Networks: Models, Methods, and Future Research , 2017, IEEE Transactions on Smart Grid.

[21]  Antonio Liotta,et al.  On-Line Building Energy Optimization Using Deep Reinforcement Learning , 2017, IEEE Transactions on Smart Grid.

[22]  Jianzhong Wu,et al.  Coordinated Control Method of Voltage and Reactive Power for Active Distribution Networks Based on Soft Open Point , 2017, IEEE Transactions on Sustainable Energy.

[23]  Yonghua Song,et al.  Model Predictive Control of LPC-Looped Active Distribution Network With High Penetration of Distributed Generation , 2017, IEEE Transactions on Sustainable Energy.

[24]  Birgitte Bak-Jensen,et al.  Predictive Control of Flexible Resources for Demand Response in Active Distribution Networks , 2019, IEEE Transactions on Power Systems.

[25]  Na Li,et al.  Optimal Distributed Feedback Voltage Control Under Limited Reactive Power , 2018, IEEE Transactions on Power Systems.

[26]  P. Dayan,et al.  States versus Rewards: Dissociable Neural Prediction Error Signals Underlying Model-Based and Model-Free Reinforcement Learning , 2010, Neuron.

[27]  Göran Andersson,et al.  Optimal Placement and Sizing of Distributed Battery Storage in Low Voltage Grids Using Receding Horizon Control Strategies , 2016, IEEE Transactions on Power Systems.

[28]  Paulo F. Ribeiro,et al.  Aiding Power System Support by Means of Voltage Control With Intelligent Distribution Substation , 2014, IEEE Transactions on Smart Grid.

[29]  Chao Long,et al.  Multi-objective operation optimization of an electrical distribution network with soft open point , 2017 .

[30]  V. Khadkikar,et al.  Enhancing Electric Power Quality Using UPQC: A Comprehensive Overview , 2012, IEEE Transactions on Power Electronics.

[31]  Kashem M. Muttaqi,et al.  Effective Utilization of Available PEV Battery Capacity for Mitigation of Solar PV Impact and Grid Support With Integrated V2G Functionality , 2016, IEEE Transactions on Smart Grid.

[32]  Johann W. Kolar,et al.  Applicability of Solid-State Transformers in Today’s and Future Distribution Grids , 2019, IEEE Transactions on Smart Grid.

[33]  N.D. Hatziargyriou,et al.  Reinforcement learning for reactive power control , 2004, IEEE Transactions on Power Systems.