Reinforcement Mechanism Design for Electric Vehicle Demand Response in Microgrid Charging Stations

Reinforcement learning has become an important scheduling solution with many successes in markets with dynamic pricing options, e.g., electric vehicle charging in a deregulated electricity market. However, the highly-uncertain requests and partially-unknown individual preferences remain major challenges to effective demand responses in the user-centric environment. For charging stations who aim to maximize the long-term revenue in this fast-growing market, an accurate estimate of user’s sensitivity, or acceptance, of the prices they offered to the potential customers is the key to the success of dynamic pricing. While most existing pricing schemes assume users will consistently follow stable patterns that are observable or inferrable by the charging service provider, it remains crucial to consider how users may be influenced by historic prices they have observed and react strategically to decide optimal charging demands that can maximize their utilities. To overcome this limitation, this paper presents a new framework based on reinforcement mechanism design to determine the optimal charging price in a mechanism design setting, which can optimize the long-term revenue of charging stations as well as the social welfare of users with private utility functions. Specifically, the strategic interaction between the station and users is modelled as a discrete finite Markov decision process, a Q-learning-based dynamic pricing mechanism is proposed to explore how price affects users’ demands over a sequence of time. The experiments demonstrate that our pricing mechanism outperforms the predetermined time-of-use pricing in maximizing the long-term revenue of the charging station.

[1]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[2]  Xiao-Jun Zeng,et al.  An optimal real-time pricing for demand-side management: A Stackelberg game and genetic algorithm approach , 2014, 2014 International Joint Conference on Neural Networks (IJCNN).

[3]  Francesco Piazza,et al.  Energy management with the support of dynamic pricing strategies in real micro-grid scenarios , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[4]  Lingfeng Wang,et al.  Optimal Day-Ahead Charging Scheduling of Electric Vehicles Through an Aggregative Game Model , 2018, IEEE Transactions on Smart Grid.

[5]  Habib Rajabi Mashhadi,et al.  An Adaptive $Q$-Learning Algorithm Developed for Agent-Based Computational Modeling of Electricity Market , 2010, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[6]  Quanyan Zhu,et al.  Dependable Demand Response Management in the Smart Grid: A Stackelberg Game Approach , 2013, IEEE Transactions on Smart Grid.

[7]  Derong Liu The Mathematics of Internet Congestion Control , 2005, IEEE Transactions on Automatic Control.

[8]  Hao Liang,et al.  Distributed Economic Dispatch in Microgrids Based on Cooperative Reinforcement Learning , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[9]  Christos V. Verikoukis,et al.  A Survey on Demand Response Programs in Smart Grids: Pricing Methods and Optimization Algorithms , 2015, IEEE Communications Surveys & Tutorials.

[10]  Maria-Florina Balcan,et al.  Mechanism design via machine learning , 2005, 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS'05).

[11]  Xiaofeng Wang,et al.  Reinforcement Learning to Play an Optimal Nash Equilibrium in Team Markov Games , 2002, NIPS.

[12]  Reza Keypour,et al.  A two-stage stochastic model for energy storage planning in a microgrid incorporating bilateral contracts and demand response program , 2019 .

[13]  Danny H. K. Tsang,et al.  Eliciting Multi-Dimensional Flexibilities From Electric Vehicles: A Mechanism Design Approach , 2019, IEEE Transactions on Power Systems.

[14]  Seung Ho Hong,et al.  Incentive-based demand response for smart grid with reinforcement learning and deep neural network , 2019, Applied Energy.

[15]  Pingzhong Tang,et al.  Reinforcement mechanism design , 2017, IJCAI.

[16]  David C. Parkes,et al.  Iterative combinatorial auctions: achieving economic and computational efficiency , 2001 .

[17]  Tuomas Sandholm,et al.  Automated Mechanism Design: A New Application Area for Search Algorithms , 2003, CP.

[18]  D. Bergemann,et al.  Dynamic Mechanism Design: An Introduction , 2017, Journal of Economic Literature.

[19]  Yan Gao,et al.  Real-time pricing scheme based on Stackelberg game in smart grid with multiple power retailers , 2017, Neurocomputing.

[20]  Haibo He,et al.  Residential Energy Management with Deep Reinforcement Learning , 2018, 2018 International Joint Conference on Neural Networks (IJCNN).

[21]  Jianwei Huang,et al.  An Online Learning Algorithm for Demand Response in Smart Grid , 2018, IEEE Transactions on Smart Grid.

[22]  Haibo He,et al.  Constrained EV Charging Scheduling Based on Safe Deep Reinforcement Learning , 2020, IEEE Transactions on Smart Grid.

[23]  José R. Vázquez-Canteli,et al.  Reinforcement learning for demand response: A review of algorithms and modeling techniques , 2019, Applied Energy.

[24]  Pravin Varaiya,et al.  Mechanism Design for Demand Response Programs , 2017, IEEE Transactions on Smart Grid.

[25]  Athanasios V. Vasilakos,et al.  Enhancing smart grid with microgrids: Challenges and opportunities , 2017 .

[26]  Mihaela van der Schaar,et al.  Dynamic Pricing and Energy Consumption Scheduling With Reinforcement Learning , 2016, IEEE Transactions on Smart Grid.