Multiagent Bayesian Deep Reinforcement Learning for Microgrid Energy Management Under Communication Failures

Microgrids (MGs) are important players for the future transactive energy systems where a number of intelligent Internet of Things (IoT) devices interact for energy management in the smart grid. Although there have been many works on MG energy management, most studies assume a perfect communication environment, where communication failures are not considered. In this paper, we consider the MG as a multiagent environment with IoT devices in which AI agents exchange information with their peers for collaboration. However, the collaboration information may be lost due to communication failures or packet loss. Such events may affect the operation of the whole MG. To this end, we propose a multi-agent Bayesian deep reinforcement learning (BA-DRL) method for MG energy management under communication failures. We first define a multi-agent partially observable Markov decision process (MAPOMDP) to describe agents under communication failures, in which each agent can update its beliefs on the actions of its peers. Then, we apply a double deep Q-learning (DDQN) architecture for Q-value estimation in BA-DRL, and propose a belief-based correlated equilibrium for the joint-action selection of multiagent BA-DRL. Finally, the simulation results show that BA-DRL is robust to both power supply uncertainty and communication failure uncertainty. BA-DRL has 4.1% and 10.3% higher reward than Nash Deep Q-learning (Nash-DQN) and alternating direction method of multipliers (ADMM) respectively under 1% communication failure probability.

[1]  Hao Liang,et al.  Distributed Economic Dispatch in Microgrids Based on Cooperative Reinforcement Learning , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Thomas Morstyn,et al.  Multiclass Energy Management for Peer-to-Peer Energy Trading Driven by Prosumer Preferences , 2019, IEEE Transactions on Power Systems.

[3]  Keith B. Hall,et al.  Correlated Q-Learning , 2003, ICML.

[4]  Jürgen Schmidhuber,et al.  Learning to Forget: Continual Prediction with LSTM , 2000, Neural Computation.

[5]  Pierluigi Siano,et al.  A Survey on Microgrid Energy Management Considering Flexible Energy Sources , 2019, Energies.

[6]  Melike Erol-Kantarci,et al.  Power Loss-Aware Transactive Microgrid Coalitions under Uncertainty , 2020 .

[7]  Xinyu Yang,et al.  A Survey on Internet of Things: Architecture, Enabling Technologies, Security and Privacy, and Applications , 2017, IEEE Internet of Things Journal.

[8]  Gang Chen,et al.  An ADMM-Based Distributed Algorithm for Economic Dispatch in Islanded Microgrids , 2018, IEEE Transactions on Industrial Informatics.

[9]  Gang Chen,et al.  Delay Effects on Consensus-Based Distributed Economic Dispatch Algorithm in Microgrid , 2018, IEEE Transactions on Power Systems.

[10]  Yu Wang,et al.  A Survey of Energy Management in Interconnected Multi-Microgrids , 2019, IEEE Access.

[11]  Wen-Zhan Song,et al.  Optimal Pricing and Energy Scheduling for Hybrid Energy Trading Market in Future Smart Grid , 2015, IEEE Transactions on Industrial Informatics.

[12]  Xi Tao,et al.  DeepSensing: A Novel Mobile Crowdsensing Framework With Double Deep Q-Network and Prioritized Experience Replay , 2020, IEEE Internet of Things Journal.

[13]  Haibo He,et al.  An Event-Driven ADR Approach for Residential Energy Resources in Microgrids With Uncertainties , 2019, IEEE Transactions on Industrial Electronics.

[14]  Mohammad Jalil Piran,et al.  A Survey on Enhanced Smart Micro-Grid Management System with Modern Wireless Technology Contribution , 2020, Energies.

[15]  Xin Dong,et al.  Energy Management Optimization of Microgrid Cluster Based on Multi-Agent-System and Hierarchical Stackelberg Game Theory , 2020, IEEE Access.

[16]  Hao Zhou,et al.  Correlated Deep Q-learning based Microgrid Energy Management , 2020, 2020 IEEE 25th International Workshop on Computer Aided Modeling and Design of Communication Links and Networks (CAMAD).

[17]  Hak-Man Kim,et al.  Double Deep $Q$ -Learning-Based Distributed Operation of Battery Energy Storage System Considering Uncertainties , 2020, IEEE Transactions on Smart Grid.

[18]  Bart De Schutter,et al.  A Comprehensive Survey of Multiagent Reinforcement Learning , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[19]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[20]  Maya R. Gupta,et al.  Introduction to the Dirichlet Distribution and Related Processes , 2010 .

[21]  Jian-Xin Xu,et al.  Consensus based approach for economic dispatch problem in a smart grid , 2013, IECON 2013 - 39th Annual Conference of the IEEE Industrial Electronics Society.

[22]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[23]  Juan C. Vasquez,et al.  Microgrid supervisory controllers and energy management systems: A literature review , 2016 .

[24]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[25]  Jhi-Young Joo,et al.  Economic Dispatch for an Agent-Based Community Microgrid , 2016, IEEE Transactions on Smart Grid.

[26]  Gabriel Garcera,et al.  Current challenges and future trends in the field of communication architectures for microgrids , 2018 .

[27]  Groupe Pdmia Markov Decision Processes In Artificial Intelligence , 2009 .

[28]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[29]  George A. Vouros,et al.  Fuzzy Q-Learning for multi-agent decentralized energy management in microgrids , 2018, Applied Energy.

[30]  Sohrab Asgarpoor,et al.  Reinforcement Learning Approach for Optimal Distributed Energy Management in a Microgrid , 2018, IEEE Transactions on Power Systems.

[31]  Yurii Nesterov,et al.  Interior-point polynomial algorithms in convex programming , 1994, Siam studies in applied mathematics.

[32]  Qiuye Sun,et al.  Distributed Optimal Economic Dispatch for Microgrids Considering Communication Delays , 2019, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[33]  Hao Zhou,et al.  Decentralized Microgrid Energy Management: A Multi-agent Correlated Q-learning Approach , 2020, 2020 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm).

[34]  Xinghuo Yu,et al.  Distributed Multi-DER Cooperative Control for Master-Slave-Organized Microgrid Networks With Limited Communication Bandwidth , 2019, IEEE Transactions on Industrial Informatics.

[35]  Melike Erol-Kantarci,et al.  Staleness Control for Edge Data Analytics , 2020, Abstracts of the 2020 SIGMETRICS/Performance Joint International Conference on Measurement and Modeling of Computer Systems.

[36]  Liang Xiao,et al.  Reinforcement Learning-Based Microgrid Energy Trading With a Reduced Power Plant Schedule , 2019, IEEE Internet of Things Journal.

[37]  Saifur Rahman,et al.  Communication network requirements for major smart grid applications in HAN, NAN and WAN , 2014, Comput. Networks.