Decentralized Planning-Assisted Deep Reinforcement Learning for Collision and Obstacle Avoidance in UAV Networks

In this paper, we propose using a decentralized planning-assisted approach of deep reinforcement learning for collision and obstacle avoidance in UAV networks. We focus on a UAV network where there are multiple UAVs and multiple static obstacles. To avoid hitting obstacles without severely deviating from the ideal UAV trajectories, we propose merging adjacent obstacles based on convex hulls and design a novel trajectory planning algorithm. For UAVs to efficiently avoid collisions in a distributed manner, we propose using a decentralized multi-agent deep reinforcement learning approach based on policy gradients. In addition, we propose using a priority-based algorithm for avoiding collisions without reducing the speeds of UAVs too much. Simulation results show that the proposed decentralized planning-assisted deep reinforcement learning approach outperforms a number of baseline approaches in terms of the probability that all UAVs successfully reach their goals within the deadline.

[1]  Qingqing Wu,et al.  Joint Trajectory and Communication Design for Multi-UAV Enabled Wireless Networks , 2017, IEEE Transactions on Wireless Communications.

[2]  Imen Mahjri,et al.  SLIDE: A Straight Line Conflict Detection and Alerting Algorithm for Multiple Unmanned Aerial Vehicles , 2018, IEEE Transactions on Mobile Computing.

[3]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[4]  Sina Sharif Mansouri,et al.  Nonlinear MPC for Collision Avoidance and Control of UAVs With Dynamic Obstacles , 2020, IEEE Robotics and Automation Letters.

[5]  Kuo-Chu Chang,et al.  UAV Path Planning with Tangent-plus-Lyapunov Vector Field Guidance and Obstacle Avoidance , 2013, IEEE Transactions on Aerospace and Electronic Systems.

[6]  Shalabh Bhatnagar,et al.  Memory-Based Deep Reinforcement Learning for Obstacle Avoidance in UAV With Limited Environment Knowledge , 2018, IEEE Transactions on Intelligent Transportation Systems.

[7]  Frank Dellaert,et al.  Path planning with uncertainty: Voronoi Uncertainty Fields , 2013, 2013 IEEE International Conference on Robotics and Automation.

[8]  Yi Wu,et al.  Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.

[9]  Yi Zhang,et al.  UAV Trajectory Design for Obstacle Avoidance Based on Cell-Varying JPS in Smart Cities , 2019, 2019 IEEE 90th Vehicular Technology Conference (VTC2019-Fall).

[10]  Tingting Sun,et al.  Flocking Control of Fixed-Wing UAVs With Cooperative Obstacle Avoidance Capability , 2019, IEEE Access.

[11]  Yuan Shen,et al.  Autonomous Navigation of UAVs in Large-Scale Complex Environments: A Deep Reinforcement Learning Approach , 2019, IEEE Transactions on Vehicular Technology.

[12]  Mariana Luderitz Kolberg,et al.  Ouroboros: Using potential field in unexplored regions to close loops , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[13]  Rung-Hung Gau,et al.  Reinforcement Learning-Based Collision Avoidance and Optimal Trajectory Planning in UAV Communication Networks , 2022, IEEE Transactions on Mobile Computing.

[14]  Rui Zhang,et al.  Wireless communications with unmanned aerial vehicles: opportunities and challenges , 2016, IEEE Communications Magazine.

[15]  Gerald Tesauro,et al.  Extending Q-Learning to General Adaptive Multi-Agent Systems , 2003, NIPS.

[16]  Srikanth Saripalli,et al.  Sampling-Based Path Planning for UAV Collision Avoidance , 2017, IEEE Transactions on Intelligent Transportation Systems.