A Deep Q Network Approach for Optimizing Offering Strategies in Electricity Markets

Bi-level optimization constitutes the most common mathematical methodology for modeling the decision-making process of strategic generation companies in deregulated electricity markets. However, previous models neglect the physical non-convex operating characteristics of generation units, due to their inherent inability to capture binary decision variables in their representation of the market clearing process, rendering them problematic in the context of markets with complex biding and unit commitment clearing mechanisms. Aiming at addressing this fundamental limitation, this paper delves into deep reinforcement learning-based approaches by proposing a novel deep Q network (DQN) method and enabling explicit incorporation of these non-convexities into the bi-level optimization model. Case studies demonstrate the superior performance of the proposed method over the conventional Q-learning method in devising more profitable offering strategies.

[1]  J. Pang,et al.  Strategic gaming analysis for electric power systems: an MPEC approach , 2000 .

[2]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[3]  Habib Rajabi Mashhadi,et al.  Measurement of Power Supplier's Market Power Using a Proposed Fuzzy Estimator , 2011, IEEE Transactions on Power Systems.

[4]  Claudio Gentile,et al.  Boltzmann Exploration Done Right , 2017, NIPS.

[5]  Goran Strbac,et al.  Investigating the Ability of Demand Shifting to Mitigate Electricity Producers’ Market Power , 2018, IEEE Transactions on Power Systems.

[6]  Enzo Sauma,et al.  Basic theoretical foundations and insights on bilevel models and their applications to power systems , 2017, Ann. Oper. Res..

[7]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[8]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[9]  T. Das,et al.  A Reinforcement Learning Model to Assess Market Power Under Auction-Based Energy Pricing , 2007, IEEE Transactions on Power Systems.

[10]  M. Carrion,et al.  A computationally efficient mixed-integer linear formulation for the thermal unit commitment problem , 2006, IEEE Transactions on Power Systems.

[11]  D. Kirschen,et al.  Fundamentals of power system economics , 1991 .

[12]  R. Srikant,et al.  Why Deep Neural Networks for Function Approximation? , 2016, ICLR.

[13]  John N. Tsitsiklis,et al.  Analysis of temporal-difference learning with function approximation , 1996, NIPS 1996.

[14]  Goran Strbac,et al.  Factoring Flexible Demand Non-Convexities in Electricity Markets , 2016, IEEE Transactions on Power Systems.

[15]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[16]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[17]  Dirk Ormoneit,et al.  Kernel-Based Reinforcement Learning , 2017, Encyclopedia of Machine Learning and Data Mining.

[18]  Haili Song,et al.  Optimal electricity supply bidding by Markov decision process , 2000 .

[19]  N. P. Ziogos,et al.  Electricity Producer Offering Strategies in Day-Ahead Energy Market With Step-Wise Offers , 2007, IEEE Transactions on Power Systems.

[20]  Morteza Rahimiyan,et al.  Supplier's optimal bidding strategy in electricity pay-as-bid auction: Comparison of the Q-learning and a model-based approach , 2008 .

[21]  A. Conejo,et al.  Pool Strategy of a Producer With Endogenous Formation of Locational Marginal Prices , 2009, IEEE Transactions on Power Systems.

[22]  Gaofeng Xiong,et al.  An electricity supplier bidding strategy through Q-Learning , 2002, IEEE Power Engineering Society Summer Meeting,.

[23]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.