A Multi-Agent System toward the Green Edge Computing with Microgrid

The nature of multi-access edge computing (MEC) is to deal with heterogeneous computational tasks near to the end users, which induces the volatile energy consumption for the MEC network. As an energy supplier, a microgrid is able to enable seamless energy flow from renewable and non- renewable sources. In particular, the risk of energy demand and supply is increased due to nondeterministic nature of both energy consumption and generation. In this paper, we impose a risk- sensitive energy profiling problem for a microgrid-enabled MEC network, where we first formulate an optimization problem by considering Conditional Value-at-Risk (CVaR). Hence, the formulated problem can determine the risk of expected energy shortfall by coordinating with the uncertainties of both demand and supply, and we show this problem is NP-hard. Second, we design a multi-agent system that can determine a risk- sensitive energy profiling by coping with an optimal scheduling policy among the agents. Third, we devise the solution by applying a multi-agent deep reinforcement learning (MADRL) based on asynchronous advantage actor-critic (A3C) algorithm with shared neural networks. This approach mitigates the curse of dimensionality for state space and also, can admit the best energy profile policy among the agents. Finally, the experimental results establish the significant performance gain of the proposed model than that a single agent solution and achieves a high accuracy energy profiling with respect to risk constraint.

[1]  Nirwan Ansari,et al.  Smart Grid Enabled Mobile Networks: Jointly Optimizing BS Operation and Power Distribution , 2017, IEEE/ACM Transactions on Networking.

[2]  Steven D. Whitehead,et al.  A Complexity Analysis of Cooperative Mechanisms in Reinforcement Learning , 1991, AAAI.

[3]  Scott Aaronson,et al.  Guest Column: NP-complete problems and physical reality , 2005, SIGA.

[4]  Choong Seon Hong,et al.  When Edge Computing Meets Microgrid: A Deep Reinforcement Learning Approach , 2019, IEEE Internet of Things Journal.

[5]  Zhu Han,et al.  User Scheduling and Resource Allocation in HetNets With Hybrid Energy Supply: An Actor-Critic Reinforcement Learning Approach , 2018, IEEE Transactions on Wireless Communications.

[6]  Yi Wu,et al.  Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.

[7]  Muhammad Ali Imran,et al.  How much energy is needed to run a wireless network? , 2011, IEEE Wireless Communications.

[8]  M. Lagoudakis The 0 – 1 Knapsack Problem An Introductory Survey , 1996 .

[9]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[10]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[11]  Eduard Ayguadé,et al.  A Systematic Methodology to Generate Decomposable and Responsive Power Models for CMPs , 2013, IEEE Transactions on Computers.

[12]  Cong Wang,et al.  CRAWDAD dataset nyupoly/video (v.2014-05-09) , 2014 .

[13]  Taskin Koçak,et al.  Smart Grid Technologies: Communication Technologies and Standards , 2011, IEEE Transactions on Industrial Informatics.

[14]  Tarik Taleb,et al.  Survey on Multi-Access Edge Computing for Internet of Things Realization , 2018, IEEE Communications Surveys & Tutorials.

[15]  Mohammad Ghavamzadeh,et al.  Algorithms for CVaR Optimization in MDPs , 2014, NIPS.

[16]  Qi Zhu,et al.  Peak-Aware Online Economic Dispatching for Microgrids , 2015, IEEE Transactions on Smart Grid.

[17]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..