Task offloading algorithm of vehicle edge computing environment based on Dueling-DQN

With the wide application of the Internet of vehicles, the rapid development of intelligent vehicles provides drivers and passengers with a good driving and riding experience. However, how to process a large amount of data messages in real time on resource-limited vehicle terminals is still a huge challenge, and it will cause great energy consumption to terminal devices. In this paper, a semi-online task distribution and offloading algorithm based on Dueling-DQN is proposed for time-varying complex vehicle environments. Firstly, the vehicle offloading system is built by the reinforcement learning, because the original optimization problem is a joint optimization problem with high complexity, and divided into the vehicle scheduling sub-problem and the vehicle computation resources optimization subproblem. We predicts different vehicle offloading behaviour, and calculates total rewards value after a series of vehicle offloading action, so as to update the vehicle offloading decisions. The simulation results show that this algorithm can improve the efficiency and energy consumption of computation tasks to a certain extent.

[1]  Geoffrey Ye Li,et al.  Deep Reinforcement Learning Based Resource Allocation for V2V Communications , 2018, IEEE Transactions on Vehicular Technology.

[2]  Wenzhong Li,et al.  Efficient Multi-User Computation Offloading for Mobile-Edge Cloud Computing , 2015, IEEE/ACM Transactions on Networking.

[3]  Sebastian Ruder,et al.  An overview of gradient descent optimization algorithms , 2016, Vestnik komp'iuternykh i informatsionnykh tekhnologii.

[4]  Shin-Ming Cheng,et al.  eNB Selection for Machine Type Communications Using Reinforcement Learning Based Markov Decision Process , 2017, IEEE Transactions on Vehicular Technology.

[5]  Feng Xia,et al.  Deep Reinforcement Learning for Vehicular Edge Computing , 2019, ACM Trans. Intell. Syst. Technol..

[6]  Xin Liu,et al.  Adaptive Learning-Based Task Offloading for Vehicular Edge Computing Systems , 2019, IEEE Transactions on Vehicular Technology.

[7]  F. Richard Yu,et al.  Distributed Optimal Relay Selection in Wireless Cooperative Networks With Finite-State Markov Channels , 2010, IEEE Transactions on Vehicular Technology.

[8]  Zhanyu Ma,et al.  Vehicular Edge Computing via Deep Reinforcement Learning , 2019, ArXiv.

[9]  Tom Schaul,et al.  Dueling Network Architectures for Deep Reinforcement Learning , 2015, ICML.

[10]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[11]  Yan Zhang,et al.  Joint Offloading and Resource Allocation in Vehicular Edge Computing and Networks , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[12]  Li Zhao,et al.  Vehicle-to-Everything (v2x) Services Supported by LTE-Based Systems and 5G , 2017, IEEE Communications Standards Magazine.

[13]  Suzhi Bi,et al.  Offloading and Resource Allocation With General Task Graph in Mobile Edge Computing: A Deep Reinforcement Learning Approach , 2020, IEEE Transactions on Wireless Communications.

[14]  Shuguang Cui,et al.  Joint offloading and computing optimization in wireless powered mobile-edge computing systems , 2017, 2017 IEEE International Conference on Communications (ICC).

[15]  Thanh Tam Nguyen,et al.  Efficient User Guidance for Validating Participatory Sensing Data , 2019, ACM Trans. Intell. Syst. Technol..

[16]  Ling Tang,et al.  Multi-User Computation Offloading in Mobile Edge Computing: A Behavioral Perspective , 2018, IEEE Network.

[17]  Geoffrey Ye Li,et al.  Spectrum Sharing in Vehicular Networks Based on Multi-Agent Reinforcement Learning , 2019, IEEE Journal on Selected Areas in Communications.