Task Offloading for UAV-based Mobile Edge Computing via Deep Reinforcement Learning

With rapid increase of data processing demands from users in mobile edge computing (MEC), the conventional mobile edge servers (MESs) are no longer capable of providing timely and effective services. Against this background, we focus on applying unmanned aerial vehicle (UAV) as an MES to provide computational task offloading services for users. In this paper, we aim at maximizing the migration throughput of user tasks with limited energy at the UAV. To be specific, we first formulate the maximization problem as a semi-Markov decision process (SMDP) without transition probability. Then we propose the deep reinforcement learning (DRL)-based scheme of maximizing user tasks migration throughput to solve the maximization problem. The scheme realizes a maximum autonomic migration throughput of users with limited UAV energy and improves quality of service (QoS) of MEC to some extent. Simulation results demonstrate that the proposed scheme is sufficient with favourable convergence.

[1]  Ryan N. Smith,et al.  Wind-energy based path planning for Unmanned Aerial Vehicles using Markov Decision Processes , 2013, 2013 IEEE International Conference on Robotics and Automation.

[2]  Tao Gao,et al.  4G UAV communication system and hovering height optimization for public safety , 2017, International Conference on e-Health Networking, Applications and Services.

[3]  Qingqing Wu,et al.  Joint Trajectory and Communication Design for Multi-UAV Enabled Wireless Networks , 2017, IEEE Transactions on Wireless Communications.

[4]  Zhang Xiaoyi,et al.  Q learning algorithm based UAV path learning and obstacle avoidence approach , 2017, 2017 36th Chinese Control Conference (CCC).

[5]  Rui Zhang,et al.  Wireless communications with unmanned aerial vehicles: opportunities and challenges , 2016, IEEE Communications Magazine.

[6]  Qingqing Wu,et al.  Securing UAV Communications via Trajectory Optimization , 2017, GLOBECOM 2017 - 2017 IEEE Global Communications Conference.

[7]  Shaolei Ren,et al.  Online Learning for Offloading and Autoscaling in Energy Harvesting Mobile Edge Computing , 2017, IEEE Transactions on Cognitive Communications and Networking.

[8]  Tony Q. S. Quek,et al.  Secure UAV communication in the presence of active eavesdropper (invited paper) , 2017, 2017 9th International Conference on Wireless Communications and Signal Processing (WCSP).

[9]  Rui Zhang,et al.  Energy-Efficient UAV Communication With Trajectory Optimization , 2016, IEEE Transactions on Wireless Communications.

[10]  Kan Zheng,et al.  A Reinforcement Learning-Based Resource Allocation Scheme for Cloud Robotics , 2018, IEEE Access.

[11]  Qingqing Wu,et al.  Energy Tradeoff in Ground-to-UAV Communication via Trajectory Design , 2017, IEEE Transactions on Vehicular Technology.

[12]  Jun Li,et al.  Probabilistic Small-Cell Caching: Performance Analysis and Optimization , 2017, IEEE Transactions on Vehicular Technology.

[13]  B. A. White,et al.  UAV optimum energy assignment using Dijkstra's Algorithm , 2007, 2007 European Control Conference (ECC).