Balancing Imbalance: On using reinforcement learning to increase stability in smart electricity grids

Over the years interest in renewable energy sources is growing and the amount of electricity provided by such resources, such as photovoltaic (PV) or wind energy is increasing. This results in a number of changes in the way that electricity is both produced and consumed in society. The supply of renewable energy sources is highly dependent on environmental changes and therefore hard to predict and adjust. As a response new methods have been proposed to shift this control to the demand side of the electricity grid by the means of Demand Side Management. Furthermore it can be seen that there is a shift from a situation in which the supply of electricity is managed by a small group of very large suppliers to a larger group of smaller suppliers. This, among others because of the increase of for example wind farms and the opportunity for residentials to generate their own power (by, for example, installing PV panels). One of the responsibilities of the larger producers is to maintain the balance between demand and supply across the electricity grid. This means that these parties should be able to ramp up or ramp down their production whenever the demand is higher or lower than expected in different regions. These adjustments happen on the so-called balancing market. Due to minimum production rates these markets are only open to larger suppliers and not to the group of smaller prosumers (parties that both consume and produce electricity). However, as this group becomes larger this group would be a great addition to the electricity markets when it comes to maintaining stability between the supply and demand in the electricity grid. One solution to add these smaller parties to the electricity grid is to bundle groups of small prosumers (a party that both produces and consumes electricity) into a cluster, controlled by a so-called aggregator. The aggregator can then offer a certain range of power within which it is able to ramp the amount of consumption (or production) up or down. However, ramping up or down too much at a given moment, might result in an imbalance in the opposite direction at a later point in time. This means that any imbalances that might be restored by the aggregator initially might come back even more extreme in the near future. In this thesis it is shown how reinforcement learning can be applied to successfully learn the boundary conditions within which an aggregator can safely ramp up or down its consumption. Furthermore Neural Fitted CACLA (NFCACLA) is proposed as a new reinforcement learning algorithm, which is a neural fitted variant on the existing CACLA algorithm.

[1]  Mohamed E. El-Hawary,et al.  The Smart Grid—State-of-the-art and future trends , 2014, 2016 Eighteenth International Middle East Power Systems Conference (MEPCON).

[2]  Henrik Madsen,et al.  Virtual Power Plants , 2014 .

[3]  U. Rieder,et al.  Markov Decision Processes , 2010 .

[4]  D. Ernst,et al.  Power systems stability control: reinforcement learning framework , 2004, IEEE Transactions on Power Systems.

[5]  Yann Boniface,et al.  Neural fitted actor-critic , 2016, ESANN.

[6]  E. Karangelos,et al.  Towards Full Integration of Demand-Side Resources in Joint Forward Energy/Reserve Electricity Markets , 2012, IEEE Transactions on Power Systems.

[7]  Zhu Han,et al.  Demand side management to reduce Peak-to-Average Ratio using game theory in smart grid , 2012, 2012 Proceedings IEEE INFOCOM Workshops.

[8]  Danny Pudjianto,et al.  Virtual power plant and system integration of distributed energy resources , 2007 .

[9]  Ganesh K. Venayagamoorthy,et al.  Potentials and promises of computational intelligence for smart grids , 2009, 2009 IEEE Power & Energy Society General Meeting.

[10]  Giorgio Rizzoni,et al.  A-ECMS: An Adaptive Algorithm for Hybrid Electric Vehicle Energy Management , 2005, Proceedings of the 44th IEEE Conference on Decision and Control.

[11]  Astrid Nieße,et al.  Supporting Renewable Power Supply Through Distributed Coordination of Energy Resources , 2015, ICT Innovations for Sustainability.

[12]  Walid Saad,et al.  Game-Theoretic Methods for the Smart Grid: An Overview of Microgrid Systems, Demand-Side Management, and Smart Grid Communications , 2012, IEEE Signal Processing Magazine.

[13]  Marco Wiering,et al.  Using continuous action spaces to solve discrete problems , 2009, 2009 International Joint Conference on Neural Networks.

[14]  Sean R Eddy,et al.  What is dynamic programming? , 2004, Nature Biotechnology.

[15]  Pierluigi Siano,et al.  Real Time Operation of Smart Grids via FCN Networks and Optimal Power Flow , 2012, IEEE Transactions on Industrial Informatics.

[16]  Vincent W. S. Wong,et al.  Autonomous Demand-Side Management Based on Game-Theoretic Energy Consumption Scheduling for the Future Smart Grid , 2010, IEEE Transactions on Smart Grid.

[17]  Martin A. Riedmiller,et al.  Reinforcement learning in feedback control , 2011, Machine Learning.

[18]  Roland Hafner Dateneffiziente selbstlernende neuronale Regler , 2009 .

[19]  P. B. Eriksen,et al.  EcoGrid EU — A prototype for European Smart Grids , 2011, 2011 IEEE Power and Energy Society General Meeting.

[20]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[21]  Jianhui Wang,et al.  Review of real-time electricity markets for integrating Distributed Energy Resources and Demand Response , 2015 .

[22]  Martin A. Riedmiller Neural Fitted Q Iteration - First Experiences with a Data Efficient Neural Reinforcement Learning Method , 2005, ECML.

[23]  F. Galiana,et al.  Demand-side reserve offers in joint energy/reserve electricity markets , 2003 .

[24]  K. Dielmann,et al.  Virtual power plants (VPP) - a new perspective for energy generation? , 2003, Proceedings of the 9th International Scientific and Practical Conference of Students, Post-graduates Modern Techniques and Technologies, 2003. MTT 2003..

[25]  J.K.A.H.P.J. Kok,et al.  The PowerMatcher: Smart Coordination for the Smart Electricity Grid , 2013 .

[26]  Damien Ernst,et al.  A quantitative analysis of the effect of flexible loads on reserve markets , 2014, 2014 Power Systems Computation Conference.

[27]  Siobhán Clarke,et al.  Decentralised Multi-Agent Reinforcement Learning for Dynamic and Uncertain Environments , 2014, ArXiv.

[28]  G. Venekamp,et al.  Dynamic pricing by scalable energy management systems — Field experiences and simulation results using PowerMatcher , 2012, 2012 IEEE Power and Energy Society General Meeting.

[29]  Barry D. Nichols Reinforcement learning in continuous state- and action-space , 2014 .

[30]  J. K. Kok,et al.  Intelligence in Electricity Networks for Embedding Renewables and Distributed Generation , 2010 .

[31]  Yi Ding,et al.  Real-Time Market Concept Architecture for EcoGrid EU—A Prototype for European Smart Grids , 2013, IEEE Transactions on Smart Grid.

[32]  Martin A. Riedmiller,et al.  Learning ( near ) Time Optimal Control for Flexible Robot Joints , 2011 .

[33]  David Hyman Gordon,et al.  Renewable Energy Resources , 1986 .

[34]  Massoud Pedram,et al.  Designing the Optimal Pricing Policy for Aggregators in the Smart Grid , 2014, 2014 Sixth Annual IEEE Green Technologies Conference.

[35]  Daniel Pérez Palomar,et al.  Demand-Side Management via Distributed Energy Generation and Storage Optimization , 2013, IEEE Transactions on Smart Grid.

[36]  P. Kotler The Prosumer Movement : a New Challenge For Marketers , 1986 .

[37]  Wolfgang Ketter,et al.  A reinforcement learning approach to autonomous decision-making in smart electricity markets , 2013, Machine Learning.

[38]  Kit Po Wong,et al.  Recent Advancements on Smart Grids in China , 2014 .

[39]  M.A. Wiering,et al.  Reinforcement Learning in Continuous Action Spaces , 2007, 2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning.

[40]  Florian Steinke,et al.  Scalable Economic Dispatch for Smart Distribution Networks , 2015, IEEE Transactions on Power Systems.

[41]  Kevin Tomsovic,et al.  A full demand response model in co-optimized energy and reserve market , 2014 .

[42]  A.L. Dimeas,et al.  Agent based control of Virtual Power Plants , 2007, 2007 International Conference on Intelligent Systems Applications to Power Systems.

[43]  Subhransu Ranjan Samantaray,et al.  Letter to the Editor: Smart Grid Initiatives in India , 2014 .

[44]  Clark W. Gellings,et al.  Demand-side management: Concepts and methods , 1993 .

[45]  Roberto Turri,et al.  Optimisation algorithm for a virtual power plant operation , 2004, 39th International Universities Power Engineering Conference, 2004. UPEC 2004..

[46]  Damien Ernst,et al.  Using approximate dynamic programming for estimating the revenues of a hydrogen-based high-capacity storage device , 2014, 2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL).

[47]  M. Negnevitsky,et al.  Novel business models for Demand Response Exchange , 2010, IEEE PES General Meeting.

[48]  C.W. Gellings,et al.  The concept of demand-side management for electric utilities , 1985, Proceedings of the IEEE.

[49]  Eric Bakker,et al.  Constrained capacity management and cost minimisation of EV-charging in a parking garage , 2013, 2013 IEEE Grenoble Conference.

[50]  Zita Vale,et al.  An integrated approach for distributed energy resource short-term scheduling in smart grids considering realistic power system simulation , 2012 .

[51]  Paul J. Werbos,et al.  Computational Intelligence for the Smart Grid-History, Challenges, and Opportunities , 2011, IEEE Computational Intelligence Magazine.

[52]  Warren B. Powell,et al.  Handbook of Learning and Approximate Dynamic Programming , 2006, IEEE Transactions on Automatic Control.

[53]  Marco Levorato,et al.  Residential Demand Response Using Reinforcement Learning , 2010, 2010 First IEEE International Conference on Smart Grid Communications.

[54]  Frits Bliek,et al.  PowerMatching City, a living lab smart grid demonstration , 2010, 2010 IEEE PES Innovative Smart Grid Technologies Conference Europe (ISGT Europe).

[55]  Peter Vrancx,et al.  Reinforcement Learning: State-of-the-Art , 2012 .

[56]  A. F. Gubina,et al.  The flexible demand influence on the joint energy and reserve markets , 2012, 2012 IEEE Power and Energy Society General Meeting.

[57]  Johan Driesen,et al.  Multiagent Charging of Electric Vehicles Respecting Distribution Transformer Loading and Voltage Limits , 2014, IEEE Transactions on Smart Grid.

[58]  Pamela MacDougall,et al.  Mitigation of wind power fluctuations by intelligent response of demand and distributed generation , 2011, 2011 2nd IEEE PES International Conference and Exhibition on Innovative Smart Grid Technologies.

[59]  Vincent Bakker,et al.  TRIANA, a control strategy for Smart Grids : forecasting, planning and real-time control , 2012 .