Online Learning Based Mobile Robot Controller Adaptation for Slip Reduction

Slip is a very common phenomena present in wheeled mobile robotic systems. It has undesirable consequences such as wasting energy and impeding system stability. To tackle the challenge of mobile robot trajectory tracking under slippery conditions, we propose a hierarchical framework that learns and adapts gains of the tracking controllers simultaneously online. Concretely, a reinforcement learning (RL) module is used to auto-tune parameters in a lateral predictive controller and a longitudinal speed PID controller. Experiments show the necessity of simultaneous gain tuning, and have demonstrated that our online framework outperforms the best baseline controller using fixed gains. By utilizing online gain adaptation, our framework achieves robust tracking performance by rejecting slip and reducing tracking errors when the mobile robot travels through various terrains.

[1]  M. Tomizuka,et al.  Reinforcement Learning Based Online Parameter Adaptation for Model Predictive Tracking Control Under Slippery Condition , 2022, 2022 American Control Conference (ACC).

[2]  Dong Yeop Kim,et al.  Climbing control of autonomous mobile robot with estimation of wheel slip and wheel-ground contact angle , 2022, Journal of Mechanical Science and Technology.

[3]  C. Karen Liu,et al.  COCOI: Contact-aware Online Context Inference for Generalizable Non-planar Pushing , 2020, 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[4]  Shady A. Maged,et al.  A path-tracking algorithm using predictive Stanley lateral controller , 2020 .

[5]  Masayoshi Tomizuka,et al.  Disturbance-Observer-Based Tracking Controller for Neural Network Driving Policy Transfer , 2020, IEEE Transactions on Intelligent Transportation Systems.

[6]  Junyong Zhai,et al.  A Trajectory Tracking Method for Wheeled Mobile Robots Based on Disturbance Observer , 2020 .

[7]  Masayoshi Tomizuka,et al.  Cascade Attribute Network: Decomposing Reinforcement Learning Control Policies using Hierarchical Neural Networks , 2020, IFAC-PapersOnLine.

[8]  L. Tai,et al.  High-Speed Autonomous Drifting With Deep Reinforcement Learning , 2020, IEEE Robotics and Automation Letters.

[9]  Mariano De Paula,et al.  Double Q-PID algorithm for mobile robot control , 2019, Expert Syst. Appl..

[10]  Pinhas Ben-Tzvi,et al.  Active Disturbance Rejection Control for Handling Slip in Tracked Vehicle Locomotion , 2019, Journal of Mechanisms and Robotics.

[11]  Masayoshi Tomizuka,et al.  Zero-shot Deep Reinforcement Learning Driving Policy Transfer for Autonomous Vehicles based on Robust Control , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[12]  Atsushi Sakai,et al.  PythonRobotics: a Python code collection of robotics algorithms , 2018, ArXiv.

[13]  Jihong Lee,et al.  A kinematic-based rough terrain control for traction and energy saving of an exploration rover , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[14]  S. Nandy,et al.  Detailed slip dynamics for nonholonomic mobile robotic system , 2011, 2011 IEEE International Conference on Mechatronics and Automation.

[15]  Yu Tian,et al.  Modeling and control of a nonholonomic Wheeled Mobile Robot with wheel slip dynamics , 2009, 2009 IEEE Symposium on Computational Intelligence in Control and Automation.