Computation Offloading in Energy Harvesting Systems via Continuous Deep Reinforcement Learning

As a promising technology to improve the computation experience for mobile devices, mobile edge computing (MEC) is becoming an emerging paradigm to meet the tremendous increasing computation demands. In this paper, a mobile edge computing system consisting of multiple mobile devices with energy harvesting and an edge server is considered. Specifically, multiple devices decide the offloading ratio and local computation capacity, which are both in continuous values. Each device equips a task load queue and energy harvesting, which increases the system dynamics and leads to the time-dependence of the optimal offloading decision. In order to minimize the sum cost of the execution time and energy consumption in the long-term, we develop a continuous control based deep reinforcement learning algorithm for computation offloading. Utilizing the actor-critic learning approach, we propose a centralized learning policy for each device. By incorporating the states of other devices with centralized learning, the proposed method learns to coordinate among all devices. Simulation results validate the effectiveness of our proposed algorithm, which demonstrates superior generalization ability and achieves a better performance compared with discrete decision based deep reinforcement learning methods.

[1]  Jun Du,et al.  Double Auction Mechanism Design for Video Caching in Heterogeneous Ultra-Dense Networks , 2019, IEEE Transactions on Wireless Communications.

[2]  Gang Qu,et al.  What is the limit of energy saving by dynamic voltage scaling? , 2001, IEEE/ACM International Conference on Computer Aided Design. ICCAD 2001. IEEE/ACM Digest of Technical Papers (Cat. No.01CH37281).

[3]  Zhijin Qin,et al.  Resource Allocation for Edge Computing in IoT Networks via Reinforcement Learning , 2019, ICC 2019 - 2019 IEEE International Conference on Communications (ICC).

[4]  Jun Du,et al.  Contract Design for Traffic Offloading and Resource Allocation in Heterogeneous Ultra-Dense Networks , 2017, IEEE Journal on Selected Areas in Communications.

[5]  Tiejun Lv,et al.  Deep reinforcement learning based computation offloading and resource allocation for MEC , 2018, 2018 IEEE Wireless Communications and Networking Conference (WCNC).

[6]  Swades De,et al.  Smart RF energy harvesting communications: challenges and opportunities , 2015, IEEE Communications Magazine.

[7]  H. Vincent Poor,et al.  Latency and Reliability-Aware Task Offloading and Resource Allocation for Mobile Edge Computing , 2017, 2017 IEEE Globecom Workshops (GC Wkshps).

[8]  Chau Yuen,et al.  Energy Efficiency Tradeoff Mechanism Towards Wireless Green Communication: A Survey , 2016, IEEE Communications Surveys & Tutorials.

[9]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[10]  Weihua Zhuang,et al.  Learning-Based Computation Offloading for IoT Devices With Energy Harvesting , 2017, IEEE Transactions on Vehicular Technology.

[11]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[12]  Mehdi Bennis,et al.  Optimized Computation Offloading Performance in Virtual Edge Computing Systems Via Deep Reinforcement Learning , 2018, IEEE Internet of Things Journal.

[13]  K. B. Letaief,et al.  A Survey on Mobile Edge Computing: The Communication Perspective , 2017, IEEE Communications Surveys & Tutorials.

[14]  Khaled Ben Letaief,et al.  Dynamic Computation Offloading for Mobile-Edge Computing With Energy Harvesting Devices , 2016, IEEE Journal on Selected Areas in Communications.