Deep Reinforcement Learning Based Approach for Online Service Placement and Computation Resource Allocation in Edge Computing

Due to the urgent emergence of computation-intensive intelligent applications on end devices, edge computing has been put forward as an extension of cloud computing, to satisfy the low-latency requirements of these applications. To process heterogenous computation tasks on an edge node, the corresponding services should be placed in advance, including installing softwares and caching databases/libraries. Considering the limited storage space and computation resources on the edge node, services should be elaborately selected and deployed on the edge node and its computation resources should be carefully allocated to placed services, according to the arrivals of computation workloads. The joint service placement and computation resource allocation problem is particularly complicated, in terms of considering the stochastic arrivals of tasks, the additional latency incurred by service migration, and the waiting time of unprocessed tasks. Benefiting from deep reinforcement learning, we propose a novel approach based on parameterized deep Q networks to make the joint service placement and computation resource allocation decisions, with the objective of minimizing the total latency of tasks in a long term. Extensive simulations are conducted to evaluate the convergence and performance achieved by our proposed approach.

[1]  Konstantinos Poularakis,et al.  Service Placement and Request Scheduling for Data-Intensive Applications in Edge Clouds , 2021, IEEE/ACM Transactions on Networking.

[2]  Suzhi Bi,et al.  Optimizing AI Service Placement and Resource Allocation in Mobile Edge Intelligence Systems , 2020, IEEE Transactions on Wireless Communications.

[3]  Wenjun Zhang,et al.  Bandit Learning-based Service Placement and Resource Allocation for Mobile Edge Computing , 2020, 2020 IEEE 31st Annual International Symposium on Personal, Indoor and Mobile Radio Communications.

[4]  Weifa Liang,et al.  Collaborate or Separate? Distributed Service Caching in Mobile Edge Clouds , 2020, IEEE INFOCOM 2020 - IEEE Conference on Computer Communications.

[5]  Shan Zhang,et al.  Cooperative Service Caching and Workload Scheduling in Mobile Edge Computing , 2020, IEEE INFOCOM 2020 - IEEE Conference on Computer Communications.

[6]  Weiqin Tong,et al.  A DQN-Based Approach for Online Service Placement in Mobile Edge Computing , 2020, CollaborateCom.

[7]  Song Guo,et al.  Resource Management at the Network Edge: A Deep Reinforcement Learning Approach , 2019, IEEE Network.

[8]  Thomas F. La Porta,et al.  Service Placement and Request Scheduling for Data-intensive Applications in Edge Clouds , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[9]  Xu Chen,et al.  Adaptive User-managed Service Placement for Mobile Edge Computing: An Online Learning Approach , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[10]  Tarik Taleb,et al.  Follow-Me Cloud: When Cloud Services Follow Mobile Users , 2019, IEEE Transactions on Cloud Computing.

[11]  Fei Xu,et al.  Winning at the Starting Line: Joint Network Selection and Service Placement for Mobile Edge Computing , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[12]  A. Tulino,et al.  Joint Service Placement and Request Routing in Multi-cell Mobile Edge Computing Networks , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[13]  Xiaohua Jia,et al.  Collaborative Service Placement for Mobile Edge Computing Applications , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[14]  Qing Wang,et al.  Parametrized Deep Q-Networks Learning: Reinforcement Learning with Discrete-Continuous Hybrid Action Space , 2018, ArXiv.

[15]  Xu Chen,et al.  Follow Me at the Edge: Mobility-Aware Dynamic Service Placement for Mobile Edge Computing , 2018, 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS).

[16]  Thomas F. La Porta,et al.  It's Hard to Share: Joint Service Placement and Request Scheduling in Edge Clouds with Sharable and Non-Sharable Resources , 2018, 2018 IEEE 38th International Conference on Distributed Computing Systems (ICDCS).

[17]  Max Mühlhäuser,et al.  Service Entity Placement for Social Virtual Reality Applications in Edge Computing , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[18]  Tarik Taleb,et al.  On Enabling 5G Automotive Systems Using Follow Me Edge-Cloud Concept , 2018, IEEE Transactions on Vehicular Technology.

[19]  Jie Xu,et al.  Joint Service Caching and Task Offloading for Mobile Edge Computing in Dense Networks , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[20]  George Pavlou,et al.  On Uncoordinated Service Placement in Edge-Clouds , 2017, 2017 IEEE International Conference on Cloud Computing Technology and Science (CloudCom).

[21]  Jie Xu,et al.  Collaborative Service Caching for Edge Computing in Dense Small Cell Networks , 2017, ArXiv.

[22]  Ting He,et al.  On the Complexity of Optimal Request Routing and Content Caching in Heterogeneous Cache Networks , 2017, IEEE/ACM Transactions on Networking.

[23]  K. B. Letaief,et al.  A Survey on Mobile Edge Computing: The Communication Perspective , 2017, IEEE Communications Surveys & Tutorials.

[24]  Osvaldo Simeone,et al.  Energy-Efficient Resource Allocation for Mobile Edge Computing-Based Augmented Reality Applications , 2016, IEEE Wireless Communications Letters.

[25]  Apollinaire Nadembega,et al.  Mobility prediction model-based service migration procedure for follow me cloud to support QoS and QoE , 2016, 2016 IEEE International Conference on Communications (ICC).

[26]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[27]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[28]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[29]  Min Chen,et al.  A Markov Decision Process-based service migration procedure for follow me cloud , 2014, 2014 IEEE International Conference on Communications (ICC).

[30]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[31]  Sujit Dey,et al.  Adaptive Mobile Cloud Computing to Enable Rich Mobile Multimedia Applications , 2013, IEEE Transactions on Multimedia.

[32]  P. Schrimpf,et al.  Dynamic Programming , 2011 .

[33]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[34]  Lin Zhong,et al.  uWave: Accelerometer-based personalized gesture recognition and its applications , 2009, 2009 IEEE International Conference on Pervasive Computing and Communications.

[35]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[36]  Long Ji Lin,et al.  Self-improving reactive agents based on reinforcement learning, planning and teaching , 1992, Machine Learning.

[37]  Haim Mendelson,et al.  Optimal Incentive-Compatible Priority Pricing for the M/M/1 Queue , 1990, Oper. Res..