Meta-Reinforcement Learning-Based Transferable Scheduling Strategy for Energy Management

In Home Energy Management System (HEMS), the scheduling of energy storage equipment and shiftable loads has been widely studied to reduce home energy costs. However, existing data-driven methods can hardly ensure the transferability amongst different tasks, such as customers with diverse preferences, appliances, and fluctuations of renewable energy in different seasons. This paper designs a transferable scheduling strategy for HEMS with different tasks utilizing a Meta-Reinforcement Learning (Meta-RL) framework, which can alleviate data dependence and massive training time for other data-driven methods. Specifically, a more practical and complete demand response scenario of HEMS is considered in the proposed Meta-RL framework, where customers with distinct electricity preferences, as well as fluctuating renewable energy in different seasons are taken into consideration. An inner level and an outer level are integrated in the proposed Meta-RL-based transferable scheduling strategy, where the inner and the outer level ensure the learning speed and appropriate initial model parameters, respectively. Moreover, Long Short-Term Memory (LSTM) is presented to extract the features from historical actions and rewards, which can overcome the challenges brought by the uncertainties of renewable energy and the customers’ loads, and enhance the robustness of scheduling strategies. A set of experiments conducted on practical data of Australia’s electricity network verify the performance of the transferable scheduling strategy.

[1]  Jinde Cao,et al.  Fuzzy H∞ Control of Discrete-Time Nonlinear Markov Jump Systems via a Novel Hybrid Reinforcement Q -Learning Method. , 2022, IEEE transactions on cybernetics.

[2]  Huijun Gao,et al.  Approximation-Free Robust Synchronization Control for Dual-Linear-Motors-Driven Systems With Uncertainties and Disturbances , 2022, IEEE Transactions on Industrial Electronics.

[3]  Yaochu Jin,et al.  Rethinking Individual Global Max in Cooperative Multi-Agent Reinforcement Learning , 2022, NeurIPS.

[4]  L. Carley,et al.  A Deep Reinforcement Learning Framework for High-Dimensional Circuit Linearization , 2022, IEEE Transactions on Circuits and Systems II: Express Briefs.

[5]  Yasin Yılmaz,et al.  Home Energy Recommendation System (HERS): A Deep Reinforcement Learning Method Based on Residents’ Feedback and Activity , 2022, IEEE Transactions on Smart Grid.

[6]  Jianrui Wang,et al.  Cooperative and Competitive Multi-Agent Systems: From Optimization to Games , 2022, IEEE/CAA Journal of Automatica Sinica.

[7]  Zongze Wu,et al.  Multistep Multiagent Reinforcement Learning for Optimal Energy Schedule Strategy of Charging Stations in Smart Grid , 2022, IEEE Transactions on Cybernetics.

[8]  Chun Sing Lai,et al.  A new deregulated demand response scheme for load over-shifting city in regulated power market , 2022, Applied Energy.

[9]  F. Qian,et al.  A Two-Level Energy Management Strategy for Multi-Microgrid Systems With Interval Prediction and Reinforcement Learning , 2022, IEEE Transactions on Circuits and Systems I: Regular Papers.

[10]  Kaiwen Zeng,et al.  Demand response considering user behaviour differences for load serving entity: A multi‐agent deep reinforcement learning approach , 2022, IET Energy Systems Integration.

[11]  Xin Jin,et al.  Quaternion-Based Attitude Synchronization With an Event-Based Communication Strategy , 2022, IEEE Transactions on Circuits and Systems I: Regular Papers.

[12]  G. Hancke,et al.  Two-Layered Hierarchical Optimization Strategy With Distributed Potential Game for Interconnected Hybrid Energy Systems , 2022, IEEE Transactions on Cybernetics.

[13]  Tongzhen Wei,et al.  Stochastic Model Predictive Control of Hybrid Energy Storage for Improving AGC Performance of Thermal Generators , 2021, IEEE Transactions on Smart Grid.

[14]  Lei Deng,et al.  Energy-Saving Optimization and Control of Autonomous Electric Vehicles With Considering Multiconstraints , 2021, IEEE Transactions on Cybernetics.

[15]  Jie Tan,et al.  Learning and Fast Adaptation for Grid Emergency Control via Deep Meta Reinforcement Learning , 2021, IEEE Transactions on Power Systems.

[16]  Aijuan Wang,et al.  DisEHPPC: Enabling Heterogeneous Privacy-Preserving Consensus-Based Scheme for Economic Dispatch in Smart Grids , 2020, IEEE Transactions on Cybernetics.

[17]  Shanlin Yang,et al.  Reinforcement learning-based scheduling strategy for energy storage in microgrid , 2022, Journal of Energy Storage.

[18]  Yoonho Park,et al.  Reinforcement Learning-Based Power Management Policy for Mobile Device Systems , 2021, IEEE Transactions on Circuits and Systems I: Regular Papers.

[19]  Weicong Kong,et al.  Benefits of Home Energy Storage Utilization: An Australian Case Study of Demand Charge Practices in Residential Sector , 2021, IEEE Transactions on Smart Grid.

[20]  Ljupco Kocarev,et al.  Data-Driven Resilient Control for Linear Discrete-Time Multi-Agent Networks Under Unconfined Cyber-Attacks , 2021, IEEE Transactions on Circuits and Systems I: Regular Papers.

[21]  Xinghuo Yu,et al.  Modeling and Control of Islanded DC Microgrid Clusters With Hierarchical Event-Triggered Consensus Algorithm , 2021, IEEE Transactions on Circuits and Systems I: Regular Papers.

[22]  Xiaojie Su,et al.  Adaptive Cooperative Terminal Sliding Mode Control for Distributed Energy Storage Systems , 2021, IEEE Transactions on Circuits and Systems I: Regular Papers.

[23]  Thillainathan Logenthiran,et al.  Optimization of Fuzzy Energy-Management System for Grid-Connected Microgrid Using NSGA-II , 2020, IEEE Transactions on Cybernetics.

[24]  João P. S. Catalão,et al.  Dynamic Economic Load Dispatch in Isolated Microgrids with Particle Swarm Optimisation considering Demand Response , 2020, 2020 55th International Universities Power Engineering Conference (UPEC).

[25]  Zhiqiang Wan,et al.  Real-Time Residential Demand Response , 2020, IEEE Transactions on Smart Grid.

[26]  Fangxing Li,et al.  Intelligent Multi-Microgrid Energy Management Based on Deep Neural Network and Model-Free Reinforcement Learning , 2020, IEEE Transactions on Smart Grid.

[27]  Donglian Qi,et al.  Intraday Residential Demand Response Scheme Based on Peer-to-Peer Energy Trading , 2020, IEEE Transactions on Industrial Informatics.

[28]  Yan Xu,et al.  A Multi-Agent Reinforcement Learning-Based Data-Driven Method for Home Energy Management , 2020, IEEE Transactions on Smart Grid.

[29]  Haibo He,et al.  A Deep Reinforcement Learning Based Approach for Home Energy Management System , 2020, 2020 IEEE Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT).

[30]  Zheng-Guang Wu,et al.  Cluster Tracking Performance Analysis of Linear Heterogeneous Multi-Agent Networks: A Complex Frequency Domain Approach , 2020, IEEE Transactions on Circuits and Systems I: Regular Papers.

[31]  Weicong Kong,et al.  Optimal Home Energy Management System With Demand Charge Tariff and Appliance Operational Dependencies , 2020, IEEE Transactions on Smart Grid.

[32]  Yulong Zou,et al.  Deep Reinforcement Learning for Smart Home Energy Management , 2019, IEEE Internet of Things Journal.

[33]  Kaveh Dehghanpour,et al.  A Learning-based Power Management for Networked Microgrids Under Incomplete Information , 2019 .

[34]  Yacine Rezgui,et al.  User Centered Neuro-Fuzzy Energy Management Through Semantic-Based Optimization , 2019, IEEE Transactions on Cybernetics.

[35]  José R. Vázquez-Canteli,et al.  Reinforcement learning for demand response: A review of algorithms and modeling techniques , 2019, Applied Energy.

[36]  Joel Z. Leibo,et al.  Prefrontal cortex as a meta-reinforcement learning system , 2018, bioRxiv.

[37]  Mike Preuss,et al.  Planning chemical syntheses with deep neural networks and symbolic AI , 2017, Nature.

[38]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[39]  Zeb Kurth-Nelson,et al.  Learning to reinforcement learn , 2016, CogSci.

[40]  Yunjian Xu,et al.  Optimal operation of energy storage with random renewable generation and AC/DC loads , 2018, 2016 IEEE 55th Conference on Decision and Control (CDC).

[41]  Mihaela van der Schaar,et al.  Dynamic Pricing and Energy Consumption Scheduling With Reinforcement Learning , 2016, IEEE Transactions on Smart Grid.

[42]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[43]  Miguel A. Ortega-Vazquez,et al.  Optimal scheduling of electric vehicle charging and vehicle-to-grid services at household level including battery degradation and price uncertainty , 2014 .

[44]  Witold Pedrycz,et al.  Granular Robust Mean-CVaR Feedstock Flow Planning for Waste-to-Energy Systems Under Integrated Uncertainty , 2014, IEEE Transactions on Cybernetics.

[45]  Alex S. Weddell,et al.  Photovoltaic Sample-and-Hold Circuit Enabling MPPT Indoors for Low-Power Systems , 2012, IEEE Transactions on Circuits and Systems I: Regular Papers.

[46]  Doron Shmilovitz,et al.  Optimal Power Management in Fueled Systems With Finite Storage Capacity , 2010, IEEE Transactions on Circuits and Systems I: Regular Papers.

[47]  Rong-Jong Wai,et al.  Grid-Connected Photovoltaic Generation System , 2008, IEEE Transactions on Circuits and Systems I: Regular Papers.

[48]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[49]  Q. Henry Wu,et al.  Decentralized nonlinear adaptive control for multimachine power systems via high-gain perturbation observer , 2004, IEEE Transactions on Circuits and Systems I: Regular Papers.

[50]  James S. Thorp,et al.  Continuum modeling of electromechanical dynamics in large-scale power systems , 2004, IEEE Transactions on Circuits and Systems I: Regular Papers.

[51]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.