Learning-Based Data Gathering for Information Freshness in UAV-Assisted IoT Networks

Unmanned aerial vehicle (UAV) has been widely deployed in efficient data collection for Internet of Things (IoT) networks. Information freshness in data collection can be characterized by the Age of Information (AoI). It is highly challenging to schedule multiple energy-constrained UAVs to improve information freshness especially when the generation instants of sensing samples are unpredictable. To deal with this issue, we leverage state-of-art reinforcement learning (RL) methods to design flight trajectories of UAVs without knowing the sampling mode each sensor node (SN) adopts. Each SN can sample the environment at periodical or random intervals. Multiple energy-constrained UAVs are dispatched to collect update packets from the SNs when flying over them. The UAV trajectory planning problem for AoI minimization is formulated as a Markov decision process (MDP). The objective is to minimize the average AoI of the SNs under the constraints of energy capacity and collision avoidance for the UAVs. Then, we propose two learning algorithms based on the Sarsa and value-decomposition network (VDN), respectively, which allow the UAVs to fulfill data collection tasks requested by the SNs. By learning directly from the environment, the Sarsa-based algorithm can approach the optimal policy asymptotically when certain conditions are satisfied. As one of the most popular multiagent deep RL methods, the VDN-based algorithm enables each UAV to make its own decision independently on its flight and data collection based on the partially observed network information. Simulation results validate the effectiveness of the proposed two learning-based algorithms compared with baseline policies.

[1]  Jie Hu,et al.  Average Age of Information in Wireless Powered Relay Aided Communication Network , 2022, IEEE Internet of Things Journal.

[2]  Lin Gao,et al.  User-Centric UAV Deployment and Content Placement in Cache-Enabled Multi-UAV Networks , 2022, IEEE Transactions on Vehicular Technology.

[3]  D. Owaki,et al.  A Survey of Sim-to-Real Transfer Techniques Applied to Reinforcement Learning for Bioinspired Robots , 2021, IEEE Transactions on Neural Networks and Learning Systems.

[4]  Holger Claussen,et al.  3D UAV Trajectory and Data Collection Optimisation Via Deep Reinforcement Learning , 2021, IEEE Transactions on Communications.

[5]  Weifa Liang,et al.  Data Collection Maximization in IoT-Sensor Networks via an Energy-Constrained UAV , 2021, IEEE Transactions on Mobile Computing.

[6]  Xiang-Gen Xia,et al.  A Survey on Millimeter-Wave Beamforming Enabled UAV Communications and Networking , 2021, IEEE Communications Surveys & Tutorials.

[7]  Huaiyu Dai,et al.  UAV-Aided Data Collection for Information Freshness in Wireless Sensor Networks , 2021, IEEE Transactions on Wireless Communications.

[8]  M. Caccamo,et al.  Multi-UAV Path Planning for Wireless Data Harvesting With Deep Reinforcement Learning , 2020, IEEE Open Journal of the Communications Society.

[9]  Yuanwei Liu,et al.  Machine Learning Empowered Trajectory and Passive Beamforming Design in UAV-RIS Wireless Networks , 2020, IEEE Journal on Selected Areas in Communications.

[10]  Pingyi Fan,et al.  AoI-Minimal Trajectory Planning and Data Collection in UAV-Assisted Wireless Powered IoT Networks , 2020, IEEE Internet of Things Journal.

[11]  Branka Vucetic,et al.  Optimizing Information Freshness in Two-Hop Status Update Systems Under a Resource Constraint , 2020, IEEE Journal on Selected Areas in Communications.

[12]  Huajin Tang,et al.  Event-Based Neuromorphic Vision for Autonomous Driving: A Paradigm Shift for Bio-Inspired Visual Sensing and Perception , 2020, IEEE Signal Processing Magazine.

[13]  Harpreet S. Dhillon,et al.  Neural Combinatorial Deep Reinforcement Learning for Age-Optimal Joint Trajectory and Scheduling Design in UAV-Assisted Networks , 2020, IEEE Journal on Selected Areas in Communications.

[14]  Stefano V. Albrecht,et al.  Benchmarking Multi-Agent Deep Reinforcement Learning Algorithms in Cooperative Tasks , 2020, NeurIPS Datasets and Benchmarks.

[15]  Huaiyu Dai,et al.  Deep Reinforcement Learning for Efficient Data Collection in UAV-Aided Internet of Things , 2020, 2020 IEEE International Conference on Communications Workshops (ICC Workshops).

[16]  Zhu Han,et al.  Data Freshness and Energy-Efficient UAV Navigation Optimization: A Deep Reinforcement Learning Approach , 2020, IEEE Transactions on Intelligent Transportation Systems.

[17]  Feng Lyu,et al.  Deep RL-based Trajectory Planning for AoI Minimization in UAV-assisted IoT , 2019, 2019 11th International Conference on Wireless Communications and Signal Processing (WCSP).

[18]  Yong Zeng,et al.  Completion Time Minimization for Multi-UAV-Enabled Data Collection , 2019, IEEE Transactions on Wireless Communications.

[19]  Li Wang,et al.  Minimizing Packet Expiration Loss With Path Planning in UAV-Assisted Data Sensing , 2019, IEEE Wireless Communications Letters.

[20]  Zhisheng Niu,et al.  Timely Status Update in Wireless Uplinks: Analytical Solutions With Asymptotic Optimality , 2019, IEEE Internet of Things Journal.

[21]  Eytan Modiano,et al.  Age Optimal Information Gathering and Dissemination on Graphs , 2019, IEEE INFOCOM 2019 - IEEE Conference on Computer Communications.

[22]  Hakim Ghazzai,et al.  An Energy Efficient Overlay Cognitive Radio Approach in UAV-Based Communication , 2018, 2018 IEEE Global Communications Conference (GLOBECOM).

[23]  Yang Yang,et al.  Energy-efficient multi-UAV coverage deployment in UAV networks: A game-theoretic framework , 2018, China Communications.

[24]  Rui Zhang,et al.  Uplink Cooperative NOMA for Cellular-Connected UAV , 2018, IEEE Journal of Selected Topics in Signal Processing.

[25]  Chi Harold Liu,et al.  Energy-Efficient UAV Control for Effective and Fair Communication Coverage: A Deep Reinforcement Learning Approach , 2018, IEEE Journal on Selected Areas in Communications.

[26]  Ming Chen,et al.  Joint Altitude, Beamwidth, Location, and Bandwidth Optimization for UAV-Enabled Communications , 2018, IEEE Communications Letters.

[27]  Zhisheng Niu,et al.  Decentralized Status Update for Age-of-Information Optimization in Wireless Multiaccess Channels , 2018, 2018 IEEE International Symposium on Information Theory (ISIT).

[28]  Harpreet S. Dhillon,et al.  Average Peak Age-of-Information Minimization in UAV-Assisted IoT Networks , 2018, IEEE Transactions on Vehicular Technology.

[29]  Jie Xu,et al.  Energy Minimization for Wireless Communication With Rotary-Wing UAV , 2018, IEEE Transactions on Wireless Communications.

[30]  Walid Saad,et al.  A Tutorial on UAVs for Wireless Networks: Applications, Challenges, and Open Problems , 2018, IEEE Communications Surveys & Tutorials.

[31]  Qingqing Wu,et al.  Energy Tradeoff in Ground-to-UAV Communication via Trajectory Design , 2017, IEEE Transactions on Vehicular Technology.

[32]  Jingxian Wu,et al.  Optimal Status Update for Age of Information Minimization With an Energy Harvesting Source , 2017, IEEE Transactions on Green Communications and Networking.

[33]  Joel Z. Leibo,et al.  Value-Decomposition Networks For Cooperative Multi-Agent Learning , 2017, ArXiv.

[34]  Walid Saad,et al.  Wireless Communication Using Unmanned Aerial Vehicles (UAVs): Optimal Transport Theory for Hover Time Optimization , 2017, IEEE Transactions on Wireless Communications.

[35]  Walid Saad,et al.  Mobile Unmanned Aerial Vehicles (UAVs) for Energy-Efficient Internet of Things Communications , 2017, IEEE Transactions on Wireless Communications.

[36]  Rui Zhang,et al.  Energy-Efficient UAV Communication With Trajectory Optimization , 2016, IEEE Transactions on Wireless Communications.

[37]  Rui Zhang,et al.  Wireless communications with unmanned aerial vehicles: opportunities and challenges , 2016, IEEE Communications Magazine.

[38]  Roy D. Yates,et al.  Update or wait: How to keep your data fresh , 2016, IEEE INFOCOM 2016 - The 35th Annual IEEE International Conference on Computer Communications.

[39]  Walid Saad,et al.  Unmanned Aerial Vehicle With Underlaid Device-to-Device Communications: Performance and Tradeoffs , 2015, IEEE Transactions on Wireless Communications.

[40]  Mahdi H. Miraz,et al.  A review on Internet of Things (IoT), Internet of Everything (IoE) and Internet of Nano Things (IoNT) , 2015, 2015 Internet Technologies and Applications (ITA).

[41]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[42]  Walid Saad,et al.  Drone Small Cells in the Clouds: Design, Deployment and Performance Analysis , 2014, 2015 IEEE Global Communications Conference (GLOBECOM).

[43]  Abbas Jamalipour,et al.  Modeling air-to-ground path loss for low altitude platforms in urban environments , 2014, 2014 IEEE Global Communications Conference.

[44]  Andrew W. Senior,et al.  Long Short-Term Memory Based Recurrent Neural Network Architectures for Large Vocabulary Speech Recognition , 2014, ArXiv.

[45]  Roy D. Yates,et al.  Real-time status: How often should one update? , 2012, 2012 Proceedings IEEE INFOCOM.

[46]  Roy D. Yates,et al.  Status updates through queues , 2012, 2012 46th Annual Conference on Information Sciences and Systems (CISS).

[47]  Tommi S. Jaakkola,et al.  Convergence Results for Single-Step On-Policy Reinforcement-Learning Algorithms , 2000, Machine Learning.

[48]  Michael I. Jordan,et al.  MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES , 1996 .

[49]  Geoffrey Y. Li,et al.  Age of Information, Latency, and Reliability in Intelligent Vehicular Networks , 2022, IEEE Network.