Information Freshness-Aware Task Offloading in Air-Ground Integrated Edge Computing Systems

This paper studies the problem of information freshness-aware task offloading in an air-ground integrated multi-access edge computing system, which is deployed by an infrastructure provider (InP). A third-party real-time application service provider provides computing services to the subscribed mobile users (MUs) with the limited communication and computation resources from the InP based on a long-term business agreement. Due to the dynamic characteristics, the interactions among the MUs are modelled by a non-cooperative stochastic game, in which the control policies are coupled and each MU aims to selfishly maximize its own expected long-term payoff. To address the Nash equilibrium solutions, we propose that each MU behaves in accordance with the local system states and conjectures, based on which the stochastic game is transformed into a single-agent Markov decision process. Moreover, we derive a novel online deep reinforcement learning (RL) scheme that adopts two separate double deep Q-networks for each MU to approximate the Q-factor and the post-decision Q-factor. Using the proposed deep RL scheme, each MU in the system is able to make decisions without a priori statistical knowledge of dynamics. Numerical experiments examine the potentials of the proposed scheme in balancing the age of information and the energy consumption.

[1]  Lingjie Duan,et al.  Economic Analysis of Unmanned Aerial Vehicle (UAV) Provided Mobile Services , 2020, IEEE Transactions on Mobile Computing.

[2]  Daniel R. Jiang,et al.  Structured Actor-Critic for Managing Public Health Points-of-Dispensing. , 2018, 1806.02490.

[3]  Roy D. Yates,et al.  The Age of Information in Networks: Moments, Distributions, and Sampling , 2018, IEEE Transactions on Information Theory.

[4]  Rui Zhang,et al.  Throughput Maximization for UAV-Enabled Mobile Relaying Systems , 2016, IEEE Transactions on Communications.

[5]  Mihaela van der Schaar,et al.  Joint Physical-Layer and System-Level Power Management for Delay-Sensitive Wireless Communications , 2013, IEEE Transactions on Mobile Computing.

[6]  Kaibin Huang,et al.  Multiuser Computation Offloading and Downloading for Edge Computing With Virtualization , 2018, IEEE Transactions on Wireless Communications.

[7]  Wenzhong Li,et al.  Efficient Multi-User Computation Offloading for Mobile-Edge Cloud Computing , 2015, IEEE/ACM Transactions on Networking.

[8]  Xianbin Cao,et al.  Efficient and Fair Network Selection for Integrated Cellular and Drone-Cell Networks , 2019, IEEE Transactions on Vehicular Technology.

[9]  Walid Saad,et al.  A Tutorial on UAVs for Wireless Networks: Applications, Challenges, and Open Problems , 2018, IEEE Communications Surveys & Tutorials.

[10]  Mugen Peng,et al.  Application of Machine Learning in Wireless Networks: Key Techniques and Open Issues , 2018, IEEE Communications Surveys & Tutorials.

[11]  Walid Saad,et al.  Ultra-Reliable Low-Latency Vehicular Networks: Taming the Age of Information Tail , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[12]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[13]  Preben E. Mogensen,et al.  Enabling Cellular Communication for Aerial Vehicles: Providing Reliability for Future Applications , 2020, IEEE Vehicular Technology Magazine.

[14]  Daniel Adelman,et al.  Relaxations of Weakly Coupled Stochastic Dynamic Programs , 2008, Oper. Res..

[15]  José A. Ramírez-Hernández,et al.  Optimization of Preventive Maintenance scheduling in semiconductor manufacturing models using a simulation-based Approximate Dynamic Programming approach , 2010, 49th IEEE Conference on Decision and Control (CDC).

[16]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[17]  Zhu Han,et al.  Learning to Entangle Radio Resources in Vehicular Communications: An Oblivious Game-Theoretic Perspective , 2019, IEEE Transactions on Vehicular Technology.

[18]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[19]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[20]  Lingjia Liu,et al.  Mobile-Edge Computing in the Sky: Energy Optimization for Air–Ground Integrated Networks , 2020, IEEE Internet of Things Journal.

[21]  Long-Ji Lin,et al.  Reinforcement learning for robots using neural networks , 1992 .

[22]  Xianfu Chen,et al.  Age of Information Aware Radio Resource Management in Vehicular Networks: A Proactive Deep Reinforcement Learning Perspective , 2019, IEEE Transactions on Wireless Communications.

[23]  Xiaofei Wang,et al.  Convergence of Edge Computing and Deep Learning: A Comprehensive Survey , 2019, IEEE Communications Surveys & Tutorials.

[24]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[25]  Roy D. Yates,et al.  The Age of Information: Real-Time Status Updating by Multiple Sources , 2016, IEEE Transactions on Information Theory.

[26]  Rui Zhang,et al.  Wireless communications with unmanned aerial vehicles: opportunities and challenges , 2016, IEEE Communications Magazine.

[27]  K. J. Liu,et al.  Dynamic Spectrum Sharing : A Game Theoretical Overview , 2022 .

[28]  Shuguang Cui,et al.  Joint offloading and computing optimization in wireless powered mobile-edge computing systems , 2017, 2017 IEEE International Conference on Communications (ICC).

[29]  Walid Saad,et al.  Deep Reinforcement Learning for Minimizing Age-of-Information in UAV-Assisted Networks , 2019, 2019 IEEE Global Communications Conference (GLOBECOM).

[30]  Sanjit Krishnan Kaul,et al.  Minimizing age of information in vehicular networks , 2011, 2011 8th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks.

[31]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[32]  Kai-Kit Wong,et al.  UAV-Assisted Relaying and Edge Computing: Scheduling and Trajectory Optimization , 2018, IEEE Transactions on Wireless Communications.

[33]  Symeon Papavassiliou,et al.  Risk-Aware Data Offloading in Multi-Server Multi-Access Edge Computing Environment , 2020, IEEE/ACM Transactions on Networking.

[34]  Tuomas Sandholm,et al.  Imperfect-Recall Abstractions with Bounds in Games , 2014, EC.

[35]  Roy D. Yates,et al.  Age-aware Scheduling for Asynchronous Arriving Jobs in Edge Applications , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS).

[36]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[37]  Mehdi Bennis,et al.  Optimized Computation Offloading Performance in Virtual Edge Computing Systems Via Deep Reinforcement Learning , 2018, IEEE Internet of Things Journal.

[38]  Markus Fiedler,et al.  A generic quantitative relationship between quality of experience and quality of service , 2010, IEEE Network.

[39]  Mehdi Bennis,et al.  Multi-Tenant Cross-Slice Resource Orchestration: A Deep Reinforcement Learning Approach , 2018, IEEE Journal on Selected Areas in Communications.

[40]  K. B. Letaief,et al.  A Survey on Mobile Edge Computing: The Communication Perspective , 2017, IEEE Communications Surveys & Tutorials.

[41]  Yi Zhou,et al.  A Smooth-Turn Mobility Model for Airborne Networks , 2013, IEEE Trans. Veh. Technol..

[42]  Abhijeet Bhorkar,et al.  An on-line learning algorithm for energy efficient delay constrained scheduling over a fading channel , 2008, IEEE Journal on Selected Areas in Communications.

[43]  A. M. Fink,et al.  Equilibrium in a stochastic $n$-person game , 1964 .

[44]  Xiao Liu,et al.  Trajectory Design and Power Control for Multi-UAV Assisted Wireless Networks: A Machine Learning Approach , 2018, IEEE Transactions on Vehicular Technology.

[45]  Dusit Niyato,et al.  Hierarchical Game-Theoretic and Reinforcement Learning Framework for Computational Offloading in UAV-Enabled Mobile Edge Computing Networks With Multiple Service Providers , 2019, IEEE Internet of Things Journal.

[46]  Thomas D. Burd,et al.  Processor design for portable systems , 1996, J. VLSI Signal Process..

[47]  Nicola Marchetti,et al.  Mobility in the Sky: Performance and Mobility Analysis for Cellular-Connected UAVs , 2019, IEEE Transactions on Communications.

[48]  H. Vincent Poor,et al.  Latency and Reliability-Aware Task Offloading and Resource Allocation for Mobile Edge Computing , 2017, 2017 IEEE Globecom Workshops (GC Wkshps).

[49]  Jie Gong,et al.  Analysis on Computation-Intensive Status Update in Mobile Edge Computing , 2020, IEEE Transactions on Vehicular Technology.

[50]  Richeng Jin,et al.  Deep PDS-Learning for Privacy-Aware Offloading in MEC-Enabled IoT , 2019, IEEE Internet of Things Journal.

[51]  Zhu Han,et al.  Wireless Resource Scheduling in Virtualized Radio Access Networks Using Stochastic Learning , 2018, IEEE Transactions on Mobile Computing.

[52]  Chao Xu,et al.  Optimizing Information Freshness in Computing-Enabled IoT Networks , 2019, IEEE Internet of Things Journal.