Online Velocity Control and Data Capture of Drones for the Internet of Things: An Onboard Deep Reinforcement Learning Approach

Applications of Unmanned Aerial Vehicles (UAVs) for data collection are a promising means to extend Internet-ofThings (IoT) networks into remote and hostile areas, and areas with no access to power supplies. Adequate design of velocity control and communication decisions of UAVs is critical to minimize the data packet losses of ground IoT nodes resulting from overflowing buffers and transmission failure. However, online velocity control and communication decision-making is challenging in UAV-enabled IoT networks, due to the lack of the up-to-date knowledge on the state of the IoT nodes, e.g., battery energy, buffer length and channel conditions, at the UAV. Current methodology using reinforcement learning complements real-time solutions to small-scale decision problems in static IoT networks. However, reinforcement learning is impractical for the online velocity control and communication decision in the UAV-enabled IoT network, due to its rapidly growing complexity (also known as the curse-of-dimensionality). This article discusses the design of onboard deep Q-network to deliver the online velocity control and communication decision of UAVs. The onboard deep Q-network can jointly determine the optimal patrol velocity of the UAV and decide the IoT node to be interrogated for data collection, thereby minimizing asymptotically the data packet loss of the IoT networks.

[1]  Tarik Taleb,et al.  Low-Altitude Unmanned Aerial Vehicles-Based Internet of Things Services: Comprehensive Survey and Future Perspectives , 2016, IEEE Internet of Things Journal.

[2]  Xiao Zhang,et al.  Fast Deployment of UAV Networks for Optimal Wireless Coverage , 2017, IEEE Transactions on Mobile Computing.

[3]  Eduardo Tovar,et al.  On-Board Deep Q-Network for UAV-Assisted Online Power Transfer and Data Collection , 2019, IEEE Transactions on Vehicular Technology.

[4]  Rong Du,et al.  The Sensable City: A Survey on the Deployment and Management for Smart City Monitoring , 2019, IEEE Communications Surveys & Tutorials.

[5]  Eduardo Tovar,et al.  Reinforcement Learning for Scheduling Wireless Powered Sensor Communications , 2019, IEEE Transactions on Green Communications and Networking.

[6]  Wenbo Liu,et al.  Design of a wind energy harvesting wireless sensor node , 2013, 2013 IEEE Third International Conference on Information Science and Technology (ICIST).

[7]  Wilfried Haensch,et al.  Solar-powering the Internet of Things , 2016, Science.

[8]  Mahesh Sooriyabandara,et al.  Low Power Wide Area Networks: An Overview , 2016, IEEE Communications Surveys & Tutorials.

[9]  Zhu Han,et al.  Taking Drones to the Next Level: Cooperative Distributed Unmanned-Aerial-Vehicular Networks for Small and Mini Drones , 2017, IEEE Vehicular Technology Magazine.

[10]  Rui Zhang,et al.  Energy-Efficient UAV Communication With Trajectory Optimization , 2016, IEEE Transactions on Wireless Communications.

[11]  Eduardo Tovar,et al.  Onboard Double Q-Learning for Airborne Data Capture in Wireless Powered IoT Networks , 2020, IEEE Networking Letters.

[12]  Wei Ni,et al.  Wireless Power Transfer and Data Collection in Wireless Sensor Networks , 2017, IEEE Transactions on Vehicular Technology.

[13]  Rose Qingyang Hu,et al.  Mobile Edge Computing in Unmanned Aerial Vehicle Networks , 2019, IEEE Wireless Communications.

[14]  Jie Xu,et al.  UAV-Enabled Wireless Power Transfer: Trajectory Design and Energy Optimization , 2017, IEEE Transactions on Wireless Communications.