A Generic Spatiotemporal Scheduling for Autonomous UAVs: A Reinforcement Learning-Based Approach

Considerable attention has been given to leverage a variety of smart city applications using unmanned aerial vehicles (UAVs). The rapid advances in artificial intelligence can empower UAVs with autonomous capabilities allowing them to learn from their surrounding environment and act accordingly without human intervention. In this paper, we propose a spatiotemporal scheduling framework for autonomous UAVs using reinforcement learning. The framework enables UAVs to autonomously determine their schedules to cover the maximum of pre-scheduled events spatially and temporally distributed in a given geographical area and over a pre-determined time horizon. The designed framework has the ability to update the planned schedules in case of unexpected emergency events. The UAVs are trained using the Q-learning (QL) algorithm to find effective scheduling plan. A customized reward function is developed to consider several constraints especially the limited battery capacity of the flying units, the time windows of events, and the delays caused by the UAV navigation between events. Numerical simulations show the behavior of the autonomous UAVs for various scenarios and corroborate the ability of QL to handle complex vehicle routing problems with several constraints. A comparison with an optimal deterministic solution is also provided to validate the performance of the learning-based solution.

[1]  CAROLINA LAGOS,et al.  An improved Particle Swarm Optimization Algorithm for the VRP with Simultaneous Pickup and Delivery and Time Windows , 2018, IEEE Latin America Transactions.

[2]  Yew-Soon Ong,et al.  City Vehicle Routing Problem (City VRP): A Review , 2015, IEEE Transactions on Intelligent Transportation Systems.

[3]  Jing Wang,et al.  Autonomous Vision-based Target Detection Using Unmanned Aerial Vehicle , 2018, 2018 IEEE 61st International Midwest Symposium on Circuits and Systems (MWSCAS).

[4]  Simon Lacroix,et al.  Planning to Monitor Wildfires with a Fleet of UAVs , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[5]  Rui Zhang,et al.  Wireless communications with unmanned aerial vehicles: opportunities and challenges , 2016, IEEE Communications Magazine.

[6]  Hakim Ghazzai,et al.  An exploratory search strategy for data routing in flying ad hoc networks , 2017, 2017 IEEE 28th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC).

[7]  Sidi-Mohammed Senouci,et al.  Network connectivity and area coverage for UAV fleet mobility model with energy constraint , 2016, 2016 IEEE Wireless Communications and Networking Conference.

[8]  I-Ming Chen,et al.  Autonomous navigation of UAV by using real-time model-based reinforcement learning , 2016, 2016 14th International Conference on Control, Automation, Robotics and Vision (ICARCV).

[9]  Vinay V Panicker,et al.  Solving a Heterogeneous Fleet Vehicle Routing Model - A practical approach , 2018, 2018 IEEE International Conference on System, Computation, Automation and Networking (ICSCA).

[10]  Angel A. Juan,et al.  Rich Vehicle Routing Problem , 2014, ACM Comput. Surv..

[11]  Kaarthik Sundar,et al.  Algorithms for Routing an Unmanned Aerial Vehicle in the Presence of Refueling Depots , 2013, IEEE Transactions on Automation Science and Engineering.

[12]  Halil Yetgin,et al.  Analysis and Optimization of Unmanned Aerial Vehicle Swarms in Logistics: An Intelligent Delivery Platform , 2019, IEEE Access.

[13]  Gregory Gutin,et al.  The traveling salesman problem , 2006, Discret. Optim..

[14]  Shinpei Kato,et al.  An Open Approach to Autonomous Vehicles , 2015, IEEE Micro.

[15]  Imad Jawhar,et al.  UAVs for smart cities: Opportunities and challenges , 2014, 2014 International Conference on Unmanned Aircraft Systems (ICUAS).

[16]  Arthur Richards,et al.  Ant Colony Optimization for routing and tasking problems for teams of UAVs , 2014, 2014 UKACC International Conference on Control (CONTROL).

[17]  Hakim Ghazzai,et al.  Q-learning based Routing Scheduling For a Multi-Task Autonomous Agent , 2019, 2019 IEEE 62nd International Midwest Symposium on Circuits and Systems (MWSCAS).

[18]  Marc G. Bellemare,et al.  Safe and Efficient Off-Policy Reinforcement Learning , 2016, NIPS.

[19]  Lihua Li,et al.  Resource Allocation and Basestation Placement in Cellular Networks With Wireless Powered UAVs , 2019, IEEE Transactions on Vehicular Technology.

[20]  Jie Xu,et al.  Energy Minimization for Wireless Communication With Rotary-Wing UAV , 2018, IEEE Transactions on Wireless Communications.

[21]  Sally I. McClean,et al.  UAV Position Estimation and Collision Avoidance Using the Extended Kalman Filter , 2013, IEEE Transactions on Vehicular Technology.

[22]  Efe Camci,et al.  Waitress quadcopter explores how to serve drinks by reinforcement learning , 2016, 2016 IEEE Region 10 Conference (TENCON).

[23]  Enrico Macii,et al.  A Case for a Battery-Aware Model of Drone Energy Consumption , 2018, 2018 IEEE International Telecommunications Energy Conference (INTELEC).

[24]  Hakim Ghazzai,et al.  Future UAV-Based ITS: A Comprehensive Scheduling Framework , 2019, IEEE Access.

[25]  Hakim Ghazzai,et al.  A Generic Spatiotemporal UAV Scheduling Framework for Multi-Event Applications , 2019, IEEE Access.

[26]  Lawrence V. Snyder,et al.  Reinforcement Learning for Solving the Vehicle Routing Problem , 2018, NeurIPS.