Deep Reinforcement Learning Method for Demand Response Management of Interruptible Load

As an important part of incentive demand response (DR), interruptible load (IL) can achieve a rapid response and improve demand side resilience. Yet, model-based optimization algorithms concerning with IL require the explicit physical or mathematical model of the system, which makes it difficult to adapt to realistic operation conditions. In this paper, a model-free deep reinforcement learning (DRL) method with dueling deep Q network (DDQN) structure is designed to optimize the DR management of IL under the time of use (TOU) tariff and variable electricity consumption patterns. The DDQN-based automatic demand response (ADR) architecture is firstly constructed, which provides a possibility for real-time application of DR. To obtain the maximum long-term profit, the DR management problem of IL is formulated as a Markov decision process (MDP), in which the state, action, and reward function are defined, respectively. The DDQN-based DRL algorithm is applied to solve this MDP for the DR strategy with maximum cumulative reward. The simulation results validate that the proposed algorithm with DDQN overcomes the noise and instability in traditional DQN, and realizes the goal of reducing both the peak load demand and the operation costs on the premise of regulating voltage to the safe limit.

[1]  M. Bollen,et al.  Real time optimal interruptible tariff mechanism incorporating utility-customer interactions , 2000 .

[2]  Thomas J. Overbye,et al.  The Commitment of Interruptible Load to Ensure Adequate System Primary Frequency Response , 2016, IEEE Transactions on Power Systems.

[3]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[4]  Lei Yang,et al.  Stochastic Optimization-Based Economic Dispatch and Interruptible Load Management With Increased Wind Penetration , 2016, IEEE Transactions on Smart Grid.

[5]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[6]  Hanchen Xu,et al.  Optimal Joint Bidding and Pricing of Profit-Seeking Load Serving Entity , 2017, IEEE Transactions on Power Systems.

[7]  Thillainathan Logenthiran,et al.  Demand Side Management in Smart Grid Using Heuristic Optimization , 2012, IEEE Transactions on Smart Grid.

[8]  Peter Palensky,et al.  Demand Side Management: Demand Response, Intelligent Energy Systems, and Smart Loads , 2011, IEEE Transactions on Industrial Informatics.

[9]  Pravin Varaiya,et al.  Markets and pricing for interruptible electric power , 1993 .

[10]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[11]  Long Zhao,et al.  The Impact of Time-of-Use (TOU) Rate Structure on Consumption Patterns of the Residential Customers , 2017, IEEE Transactions on Industry Applications.

[12]  Tom Schaul,et al.  Dueling Network Architectures for Deep Reinforcement Learning , 2015, ICML.

[13]  Hanchen Xu,et al.  Deep Reinforcement Learning for Joint Bidding and Pricing of Load Serving Entity , 2019, IEEE Transactions on Smart Grid.

[14]  Seung Ho Hong,et al.  A Dynamic pricing demand response algorithm for smart grid: Reinforcement learning approach , 2018, Applied Energy.

[15]  José R. Vázquez-Canteli,et al.  Reinforcement learning for demand response: A review of algorithms and modeling techniques , 2019, Applied Energy.

[16]  Jianxue Wang,et al.  The Forward Contract Model of Interruptible Load in Power Market , 2005, 2005 IEEE/PES Transmission & Distribution Conference & Exposition: Asia and Pacific.

[17]  Shahram Jadid,et al.  Multi‐objective bi‐level optimisation to design real‐time pricing for demand response programs in retail markets , 2019, IET Generation, Transmission & Distribution.

[18]  Osvaldo R. Saavedra,et al.  Decision Making in Emergency Operation for Power Transformers With Regard to Risks and Interruptible Load Contracts , 2018, IEEE Transactions on Power Delivery.

[19]  Xinghuo Yu,et al.  Efficient Computation for Sparse Load Shifting in Demand Side Management , 2017, IEEE Transactions on Smart Grid.

[20]  Christos V. Verikoukis,et al.  A Survey on Demand Response Programs in Smart Grids: Pricing Methods and Optimization Algorithms , 2015, IEEE Communications Surveys & Tutorials.

[21]  Majid Taheri Andani,et al.  Effect of Changes in Incentives and Penalties on Interruptible/Curtailable Demand Response Program in Microgrid Operation , 2019, 2019 IEEE Texas Power and Energy Conference (TPEC).

[22]  Chenming Li,et al.  Energy Management Strategy for a Hybrid Electric Vehicle Based on Deep Reinforcement Learning , 2018 .

[23]  Vincent W. S. Wong,et al.  Autonomous Demand-Side Management Based on Game-Theoretic Energy Consumption Scheduling for the Future Smart Grid , 2010, IEEE Transactions on Smart Grid.

[24]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[25]  Seung Ho Hong,et al.  Demand Response Management for Industrial Facilities: A Deep Reinforcement Learning Approach , 2019, IEEE Access.