Application of two promising Reinforcement Learning algorithms for load shifting in a cooling supply system

Abstract With the increasing use of volatile renewable energies, the requirements for building automation and control systems (BACS) are increasing. Load shifting within local energy systems stabilizes fluctuations in the grid and can be triggered by price signals. The energy purchase can thus be considered and solved as an optimal control problem. Classical approaches, often based on the optimization of mathematical models, are uneconomical in many cases, due to the high effort involved in the model creation. Algorithms from the field of Reinforcement Learning (RL), on the other hand, have a high potential for the automation of energy system optimization, due to their model-free and data-driven characteristics. However, there is still a lack of studies that examine algorithms for BACS-related applications in a structured way. Therefore, we present a study, investigating the potential of two different RL algorithms for load shifting in a cooling supply system. We combine the benefits of Modelica, a powerful modeling language, with RL algorithms and demonstrate how generalized relationships and control decisions can be learned. The case study is modeled according to a cooling supply system in Berlin, Germany. The two different algorithms (DQN and DDPG) are used to control the operation parameters of a central compression chiller, with respect to a price signal. While real monitoring data are used as exogenous influences, the thermal dynamics of the cooling network are simulated. With the learned policies, flexibility in the network is used which leads on average to weekly cost savings of 14 %, compared to direct load coverage. Our results suggest that, under certain conditions, RL is a suitable alternative to established methods. However, we also acknowledge that there are still research questions to address before RL can be applied in real BACS.

[1]  Jianwei Huang,et al.  An Online Learning Algorithm for Demand Response in Smart Grid , 2018, IEEE Transactions on Smart Grid.

[2]  Zhe Wang,et al.  Reinforcement learning for building controls: The opportunities and challenges , 2020, Applied Energy.

[3]  Jinjun Xiong,et al.  Demand-Side Management of Domestic Electric Water Heaters Using Approximate Dynamic Programming , 2017, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[4]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[5]  Farrokh Janabi-Sharifi,et al.  Theory and applications of HVAC control systems – A review of model predictive control (MPC) , 2014 .

[6]  Tom Holvoet,et al.  Reinforcement Learning of Heuristic EV Fleet Charging in a Day-Ahead Electricity Market , 2015, IEEE Transactions on Smart Grid.

[7]  Peter Palensky,et al.  Demand Side Management: Demand Response, Intelligent Energy Systems, and Smart Loads , 2011, IEEE Transactions on Industrial Informatics.

[8]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[9]  Jim Dowling,et al.  Decentralized Reinforcement Learning for the Online Optimization of Distributed Systems , 2008 .

[10]  Lei Yang,et al.  Reinforcement learning for optimal control of low exergy buildings , 2015 .

[11]  José R. Vázquez-Canteli,et al.  Balancing comfort and energy consumption of a heat pump using batch reinforcement learning with fitted Q-iteration , 2017 .

[12]  Abdul Salam Shah,et al.  A Review on Energy Consumption Optimization Techniques in IoT Based Smart Building Environments , 2019, Inf..

[13]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[14]  Xiaoqing Han,et al.  Review on the research and practice of deep learning and reinforcement learning in smart grids , 2018, CSEE Journal of Power and Energy Systems.

[15]  Michael Wooldridge,et al.  Intelligent Agents: The Key Concepts , 2001, Multi-Agent-Systems and Applications.

[16]  R. Bellman Dynamic programming. , 1957, Science.

[17]  José R. Vázquez-Canteli,et al.  Fusing TensorFlow with building energy simulation for intelligent energy management in smart cities , 2019, Sustainable Cities and Society.

[18]  José R. Vázquez-Canteli,et al.  Reinforcement learning for demand response: A review of algorithms and modeling techniques , 2019, Applied Energy.

[19]  Dirk Müller,et al.  AixLib - An Open-Source Modelica Library within the IEA-EBC Annex60 Framework , 2016 .

[20]  R. Belmans,et al.  Reinforcement Learning Applied to an Electric Water Heater: From Theory to Practice , 2015, IEEE Transactions on Smart Grid.

[21]  Biao Huang,et al.  A Long-Short Term Memory Recurrent Neural Network Based Reinforcement Learning Controller for Office Heating Ventilation and Air Conditioning Systems , 2017 .

[22]  Leslie K. Norford,et al.  Optimal control of HVAC and window systems for natural ventilation through reinforcement learning , 2018, Energy and Buildings.

[23]  George A. Vouros,et al.  Fuzzy Q-Learning for multi-agent decentralized energy management in microgrids , 2018, Applied Energy.

[24]  Johan Driesen,et al.  Deep Reinforcement Learning for Optimal Control of Space Heating , 2018, ArXiv.

[25]  Manfred Morari,et al.  Model predictive control: Theory and practice - A survey , 1989, Autom..

[26]  Lion Hirth,et al.  The ENTSO-E Transparency Platform – A review of Europe’s most ambitious electricity data platform , 2018, Applied Energy.

[27]  Yonggang Wen,et al.  Transforming Cooling Optimization for Green Data Center via Deep Reinforcement Learning , 2017, IEEE Transactions on Cybernetics.

[28]  Demis Hassabis,et al.  Mastering the game of Go without human knowledge , 2017, Nature.

[29]  Dae-Man Han,et al.  Design and implementation of smart home energy management systems based on zigbee , 2010, IEEE Transactions on Consumer Electronics.