Design and Experimental Validation of Deep Reinforcement Learning-Based Fast Trajectory Planning and Control for Mobile Robot in Unknown Environment.

This article is concerned with the problem of planning optimal maneuver trajectories and guiding the mobile robot toward target positions in uncertain environments for exploration purposes. A hierarchical deep learning-based control framework is proposed which consists of an upper level motion planning layer and a lower level waypoint tracking layer. In the motion planning phase, a recurrent deep neural network (RDNN)-based algorithm is adopted to predict the optimal maneuver profiles for the mobile robot. This approach is built upon a recently proposed idea of using deep neural networks (DNNs) to approximate the optimal motion trajectories, which has been validated that a fast approximation performance can be achieved. To further enhance the network prediction performance, a recurrent network model capable of fully exploiting the inherent relationship between preoptimized system state and control pairs is advocated. In the lower level, a deep reinforcement learning (DRL)-based collision-free control algorithm is established to achieve the waypoint tracking task in an uncertain environment (e.g., the existence of unexpected obstacles). Since this approach allows the control policy to directly learn from human demonstration data, the time required by the training process can be significantly reduced. Moreover, a noisy prioritized experience replay (PER) algorithm is proposed to improve the exploring rate of control policy. The effectiveness of applying the proposed deep learning-based control is validated by executing a number of simulation and experimental case studies. The simulation result shows that the proposed DRL method outperforms the vanilla PER algorithm in terms of training speed. Experimental videos are also uploaded, and the corresponding results confirm that the proposed strategy is able to fulfill the autonomous exploration mission with improved motion planning performance, enhanced collision avoidance ability, and less training time.

[1]  Yuanqing Xia,et al.  Design and Implementation of Deep Neural Network-Based Control for Automatic Parking Maneuver Process , 2020, IEEE Transactions on Neural Networks and Learning Systems.

[2]  Barry Lennox,et al.  Accelerated Sim-to-Real Deep Reinforcement Learning: Learning Collision Avoidance from Human Player , 2021, 2021 IEEE/SICE International Symposium on System Integration (SII).

[3]  Xin Zhou,et al.  EGO-Planner: An ESDF-Free Gradient-Based Local Planner for Quadrotors , 2020, IEEE Robotics and Automation Letters.

[4]  Fei Gao,et al.  RAPTOR: Robust and Perception-Aware Trajectory Replanning for Quadrotor Fast Flight , 2020, IEEE Transactions on Robotics.

[5]  Xiaojing Zhang,et al.  Optimization-Based Collision Avoidance , 2017, IEEE Transactions on Control Systems Technology.

[6]  E. Lucet,et al.  A modified Hybrid Reciprocal Velocity Obstacles approach for multi-robot motion planning without communication , 2020, 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[7]  Shuzhi Sam Ge,et al.  Road-Constrained Geometric Pose Estimation for Ground Vehicles , 2020, IEEE Transactions on Automation Science and Engineering.

[8]  Yuanqing Xia,et al.  Six-DOF Spacecraft Optimal Trajectory Planning and Real-Time Attitude Control: A Deep Neural Network-Based Approach , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[9]  Fei Gao,et al.  Robust and Efficient Quadrotor Trajectory Generation for Fast Autonomous Flight , 2019, IEEE Robotics and Automation Letters.

[10]  Yuanqing Xia,et al.  Two-Stage Trajectory Optimization for Autonomous Ground Vehicles Parking Maneuver , 2019, IEEE Transactions on Industrial Informatics.

[11]  Qinru Qiu,et al.  Autonomous Waypoint Planning, Optimal Trajectory Generation and Nonlinear Tracking Control for Multi-rotor UAVs , 2019, 2019 18th European Control Conference (ECC).

[12]  Frank Allgöwer,et al.  Collision avoidance for uncertain nonlinear systems with moving obstacles using robust Model Predictive Control , 2019, 2019 18th European Control Conference (ECC).

[13]  Youxian Sun,et al.  Robust State/Output-Feedback Control of Coaxial-Rotor MAVs Based on Adaptive NN Approach , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[14]  2019 International Conference on Robotics and Automation (ICRA) , 2019 .

[15]  Qinru Qiu,et al.  Autonomous waypoints planning and trajectory generation for multi-rotor UAVs , 2019, DESTION@CPSIoTWeek.

[16]  Tieshan Li,et al.  Bounded Neural Network Control for Target Tracking of Underactuated Autonomous Surface Vehicles in the Presence of Uncertain Target Dynamics , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[17]  George P. Kontoudis,et al.  Kinodynamic Motion Planning With Continuous-Time Q-Learning: An Online, Model-Free, and Safe Navigation Framework , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[18]  Zhidong Deng,et al.  A Novel Dual Successive Projection-Based Model-Free Adaptive Control Method and Application to an Autonomous Car , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Sergey Levine,et al.  Generalization through Simulation: Integrating Simulated and Real Data into Deep Reinforcement Learning for Vision-Based Autonomous Flight , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[20]  Chao Xu,et al.  Hierarchical Decision and Control for Continuous Multitarget Problem: Policy Evaluation With Action Delay , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[21]  Li Li,et al.  A Gridmap-Path Reshaping Algorithm for Path Planning , 2019, IEEE Access.

[22]  Shiji Song,et al.  Plume Tracing via Model-Free Reinforcement Learning Method , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[23]  Chen Zhang,et al.  Probabilistic Anticipation and Control in Autonomous Car Following , 2019, IEEE Transactions on Control Systems Technology.

[24]  Long Chen,et al.  A Fast and Efficient Double-Tree RRT$^*$-Like Sampling-Based Planner Applying on Mobile Robotic Systems , 2018, IEEE/ASME Transactions on Mechatronics.

[25]  Karl Henrik Johansson,et al.  Robust MPC for tracking constrained unicycle robots with additive disturbances , 2018, Autom..

[26]  Zhijun Li,et al.  Neural Network Approximation Based Near-Optimal Motion Planning With Kinodynamic Constraints Using RRT , 2018, IEEE Transactions on Industrial Electronics.

[27]  Fei-Yue Wang,et al.  Capturing Car-Following Behaviors by Deep Learning , 2018, IEEE Transactions on Intelligent Transportation Systems.

[28]  Ming Liu,et al.  Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[29]  Qian Wang,et al.  Predictive Guidance and Control Framework for (Semi-)Autonomous Vehicles in Public Traffic , 2017, IEEE Transactions on Control Systems Technology.

[30]  Haibo He,et al.  Near-Optimal Tracking Control of Mobile Robots Via Receding-Horizon Dual Heuristic Programming , 2016, IEEE Transactions on Cybernetics.

[31]  Zhijiang Shao,et al.  Time-Optimal Maneuver Planning in Automatic Parallel Parking Using a Simultaneous Dynamic Optimization Approach , 2016, IEEE Transactions on Intelligent Transportation Systems.

[32]  Yuanqing Xia,et al.  Receding horizon tracking control of unicycle‐type robots based on virtual structure , 2016 .

[33]  Xin Xu,et al.  A hierarchical path planning approach based on A⁎ and least-squares policy iteration for mobile robots , 2015, Neurocomputing.

[34]  Myoungho Sunwoo,et al.  Development of Autonomous Car—Part II: A Case Study on the Implementation of an Autonomous Driving System Based on Distributed Architecture , 2015, IEEE Transactions on Industrial Electronics.

[35]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[36]  Myoungho Sunwoo,et al.  Development of Autonomous Car—Part I: Distributed System Architecture and Development Process , 2014, IEEE Transactions on Industrial Electronics.

[37]  Chenguang Yang,et al.  Neural Network-Based Motion Control of an Underactuated Wheeled Inverted Pendulum Model , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[38]  Lorenz T. Biegler,et al.  On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming , 2006, Math. Program..