Batch-Constrained Reinforcement Learning for Dynamic Distribution Network Reconfiguration

Dynamic distribution network reconfiguration (DNR) algorithms perform hourly status changes of remotely controllable switches to improve distribution system performance. The problem is typically solved by physical model-based control algorithms, which not only rely on accurate network parameters but also lack scalability. To address these limitations, this paper develops a data-driven batch-constrained reinforcement learning (RL) algorithm for the dynamic DNR problem. The proposed RL algorithm learns the network reconfiguration control policy from a finite historical operational dataset without interacting with the distribution network. The numerical study results on three distribution networks show that the proposed algorithm not only outperforms state-of-the-art RL algorithms but also improves the behavior control policy, which generated the historical operational data. The proposed algorithm is also very scalable and can find a desirable network reconfiguration solution in real-time.

[1]  Hsiao-Dong Chiang,et al.  Toward Optimal Multiperiod Network Reconfiguration for Increasing the Hosting Capacity of Distribution Networks , 2018, IEEE Transactions on Power Delivery.

[2]  Mato Baotić,et al.  Dynamic Reconfiguration of Electrical Power Distribution Systems with Distributed Generation and Storage , 2015 .

[3]  Jiaqiao Hu,et al.  A rolling horizon approach to distribution feeder reconfiguration with switching costs , 2011, 2011 IEEE International Conference on Smart Grid Communications (SmartGridComm).

[4]  Sanjay Mehrotra,et al.  Robust Distribution Network Reconfiguration , 2015, IEEE Transactions on Smart Grid.

[5]  Dale Schuurmans,et al.  Bridging the Gap Between Value and Policy Based Reinforcement Learning , 2017, NIPS.

[6]  Luciane Neves Canha,et al.  Real‐Time Reconfiguration of Distribution Network with Distributed Generation , 2014 .

[7]  Salem Arif,et al.  Optimum dynamic distribution network reconfiguration using minimum spanning tree algorithm , 2017, 2017 5th International Conference on Electrical Engineering - Boumerdes (ICEE-B).

[8]  Matti Lehtonen,et al.  Value of Distribution Network Reconfiguration in Presence of Renewable Energy Resources , 2016, IEEE Transactions on Power Systems.

[9]  Florin Capitanescu,et al.  A Comprehensive Centralized Approach for Voltage Constraints Management in Active Distribution Grid , 2014, IEEE Transactions on Power Systems.

[10]  Zhe Chen,et al.  Comprehensive Cost Minimization in Distribution Networks Using Segmented-Time Feeder Reconfiguration and Reactive Power Control of Distributed Generators , 2016, IEEE Transactions on Power Systems.

[11]  Magdy M. A. Salama,et al.  Energy Management of AC–DC Hybrid Distribution Systems Considering Network Reconfiguration , 2019, IEEE Transactions on Power Systems.

[12]  Luis F. Ochoa,et al.  Assessing the Potential of Network Reconfiguration to Improve Distributed Generation Hosting Capacity in Active Distribution Systems , 2015, IEEE Transactions on Power Systems.

[13]  Wei Wang,et al.  Volt-VAR Control in Power Distribution Systems with Deep Reinforcement Learning , 2019, 2019 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm).

[14]  Amin Khodaei,et al.  Effective Dynamic Scheduling of Reconfigurable Microgrids , 2018, IEEE Transactions on Power Systems.

[15]  R. Jabr,et al.  Minimum Loss Network Reconfiguration Using Mixed-Integer Convex Programming , 2012, IEEE Transactions on Power Systems.

[16]  Javier García,et al.  A comprehensive survey on safe reinforcement learning , 2015, J. Mach. Learn. Res..

[17]  Felix F. Wu,et al.  Network Reconfiguration in Distribution Systems for Loss Reduction and Load Balancing , 1989, IEEE Power Engineering Review.

[18]  Zifa Liu,et al.  Intra-Day Dynamic Network Reconfiguration Based on Probability Analysis Considering the Deployment of Remote Control Switches , 2019, IEEE Access.

[19]  J. J. Grainger,et al.  Distribution feeder reconfiguration for loss reduction , 1988 .

[20]  Juan Li,et al.  Phase Identification in Electric Power Distribution Systems by Clustering of Smart Meter Data , 2016, 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA).

[21]  Herke van Hoof,et al.  Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.

[22]  Zhengcai Fu,et al.  An improved TS algorithm for loss-minimum reconfiguration in large-scale distribution systems , 2007 .

[23]  Honglak Lee,et al.  Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.

[24]  Sergey Levine,et al.  Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.

[25]  Martin A. Riedmiller,et al.  Batch Reinforcement Learning , 2012, Reinforcement Learning.

[26]  C. Su,et al.  Network Reconfiguration of Distribution Systems Using Improved Mixed-Integer Hybrid Differential Evolution , 2002, IEEE Power Engineering Review.

[27]  Hossein Haghighat,et al.  Distribution System Reconfiguration Under Uncertain Load and Renewable Generation , 2016, IEEE Transactions on Power Systems.

[28]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[29]  J. Lofberg,et al.  YALMIP : a toolbox for modeling and optimization in MATLAB , 2004, 2004 IEEE International Conference on Robotics and Automation (IEEE Cat. No.04CH37508).

[30]  D. Z. Fitiwi,et al.  Dynamic reconfiguration of distribution network systems: A key flexibility option for RES integration , 2017, 2017 IEEE International Conference on Environment and Electrical Engineering and 2017 IEEE Industrial and Commercial Power Systems Europe (EEEIC / I&CPS Europe).

[31]  Wei Wang,et al.  Dynamic Distribution Network Reconfiguration Using Reinforcement Learning , 2019, 2019 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm).

[32]  Feng Qiu,et al.  Identification of Critical Switches for Integrating Renewable Distributed Generation by Dynamic Network Reconfiguration , 2018, IEEE Transactions on Sustainable Energy.

[33]  Nanpeng Yu,et al.  A Physically Inspired Data-Driven Model for Electricity Theft Detection With Smart Meter Data , 2019, IEEE Transactions on Industrial Informatics.

[34]  Doina Precup,et al.  Off-Policy Deep Reinforcement Learning without Exploration , 2018, ICML.

[35]  Brandon Foggo,et al.  Improving Supervised Phase Identification Through the Theory of Information Losses , 2019, IEEE Transactions on Smart Grid.

[36]  D. Das A fuzzy multiobjective approach for network reconfiguration of distribution systems , 2006, IEEE Transactions on Power Delivery.

[37]  Meysam Doostizadeh,et al.  Optimal Reconfiguration of Distribution Network Using $\mu$ PMU Measurements: A Data-Driven Stochastic Robust Optimization , 2020, IEEE Transactions on Smart Grid.

[38]  Pieter Abbeel,et al.  Constrained Policy Optimization , 2017, ICML.

[39]  Wei Wang,et al.  Safe Off-Policy Deep Reinforcement Learning Algorithm for Volt-VAR Control in Power Distribution Systems , 2020, IEEE Transactions on Smart Grid.

[40]  Vahid Vahidinasab,et al.  A Resilience-Based Architecture for Joint Distributed Energy Resources Allocation and Hourly Network Reconfiguration , 2019, IEEE Transactions on Industrial Informatics.

[41]  Nanpeng Yu,et al.  A Comprehensive Evaluation of Supervised Machine Learning for the Phase Identification Problem , 2018 .

[42]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.