Reinforcement Learning Based Computation-aware Mobility Management in Ultra Dense Networks

Computation-aware delay optimal mobility management (MM) is an important problem in ultra-dense network (UDN) with mobile edge computing (MEC). Since the additional time delay caused by task computation is not taken into consideration, traditional radio access-oriented mobility management scheme cannot guarantee the overall delay performance of delay-sensitive user equipment (UE). In this paper, we propose a novel dynamic programming-based mobility management (DPMM) scheme to minimize the average delay under an energy consumption constraint. DPMM makes MM decisions using statistic information to handle the inaccurate state information. Cooperative data transmission is adopted to improve the delay performance. Simulation shows that the proposed DPMM scheme can achieve delay performance close to optimal and reduce the frequency of handover. However, the wireless link, computation resources and UE’s location in UDN environment is dynamic, which leads to information uncertainties. We further propose an MM scheme based on deep Q-network (DQN) to learn the system information from the environment. In this scheme, UE takes the current and past observed delay as experience, learning the optimal mobility management strategy through DQN training. Simulation shows that DQN-based MM can learn from experience and reduce the handover frequency to a certain degree.

[1]  Bin Li,et al.  UAV Communications for 5G and Beyond: Recent Advances and Future Trends , 2019, IEEE Internet of Things Journal.

[2]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[3]  Zhenyu Zhou,et al.  Networked MIMO With Fractional Joint Transmission in Energy Harvesting Systems , 2016, IEEE Transactions on Communications.

[4]  Rui Li,et al.  Context-Aware QoS Prediction With Neural Collaborative Filtering for Internet-of-Things Services , 2020, IEEE Internet of Things Journal.

[5]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[6]  John B. Shoven,et al.  I , Edinburgh Medical and Surgical Journal.

[7]  Ismail Güvenç,et al.  Context-aware mobility management in HetNets: A reinforcement learning approach , 2015, 2015 IEEE Wireless Communications and Networking Conference (WCNC).

[8]  Honghao Gao,et al.  V2VR: Reliable Hybrid-Network-Oriented V2V Data Transmission and Routing Considering RSUs and Connectivity Probability , 2020, IEEE Transactions on Intelligent Transportation Systems.

[9]  Jie Xu,et al.  E2M2: Energy efficient mobility management in dense small cells with mobile edge computing , 2017, 2017 IEEE International Conference on Communications (ICC).

[10]  Tsuyoshi Murata,et al.  {m , 1934, ACML.

[11]  Mengshi Hu,et al.  A Universal Predictive Mobility Management Scheme for Urban Ultra-Dense Networks With Control/Data Plane Separation , 2017, IEEE Access.

[12]  Xiaohu Ge,et al.  User Mobility Evaluation for 5G Small Cell Networks Based on Individual Mobility Model , 2015, IEEE Journal on Selected Areas in Communications.

[13]  Jie Xu,et al.  EMM: Energy-Aware Mobility Management for Mobile Edge Computing in Ultra Dense Networks , 2017, IEEE Journal on Selected Areas in Communications.

[14]  Jianping Pan,et al.  Learning Based Mobility Management Under Uncertainties for Mobile Edge Computing , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[15]  Kelvin Dias,et al.  Supporting mobility-aware computational offloading in mobile cloud environment , 2017, J. Netw. Comput. Appl..

[16]  David Tse,et al.  Fundamentals of Wireless Communication , 2005 .

[17]  Zhifeng Zhao,et al.  Deep Learning-Based Intelligent Dual Connectivity for Mobility Management in Dense Network , 2018, 2018 IEEE 88th Vehicular Technology Conference (VTC-Fall).

[18]  Goran Strbac,et al.  A Deep Q Network Approach for Optimizing Offering Strategies in Electricity Markets , 2019, 2019 International Conference on Smart Energy Systems and Technologies (SEST).

[19]  Honghao Gao,et al.  An IoT-based task scheduling optimization scheme considering the deadline and cost-aware scientific workflow for cloud computing , 2019, EURASIP Journal on Wireless Communications and Networking.

[20]  K. B. Letaief,et al.  A Survey on Mobile Edge Computing: The Communication Perspective , 2017, IEEE Communications Surveys & Tutorials.

[21]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[22]  Cong Shen,et al.  A Non-Stochastic Learning Approach to Energy Efficient Mobility Management , 2016, IEEE Journal on Selected Areas in Communications.

[23]  Holger Claussen,et al.  Towards 1 Gbps/UE in Cellular Systems: Understanding Ultra-Dense Small Cell Deployments , 2015, IEEE Communications Surveys & Tutorials.

[24]  Mohsen Guizani,et al.  Secure UAV Communication Networks over 5G , 2019, IEEE Wireless Communications.

[25]  Bin Guo,et al.  Mining consuming Behaviors with Temporal Evolution for Personalized Recommendation in Mobile Marketing Apps , 2020, Mobile Networks and Applications.

[26]  Rose Qingyang Hu,et al.  Fast and Efficient Radio Resource Allocation in Dynamic Ultra-Dense Heterogeneous Networks , 2017, IEEE Access.

[27]  Amr M. Youssef,et al.  Ultra-Dense Networks: A Survey , 2016, IEEE Communications Surveys & Tutorials.