Adaptive Tensegrity Locomotion on Rough Terrain via Reinforcement Learning

The dynamical properties of tensegrity robots give them appealing ruggedness and adaptability, but present major challenges with respect to locomotion control. Due to high-dimensionality and complex contact responses, data-driven approaches are apt for producing viable feedback policies. Guided Policy Search (GPS), a sample-efficient and model-free hybrid framework for optimization and reinforcement learning, has recently been used to produce periodic locomotion for a spherical 6-bar tensegrity robot on flat or slightly varied surfaces. This work provides an extension to non-periodic locomotion and achieves rough terrain traversal, which requires more broadly varied, adaptive, and non-periodic rover behavior. The contribution alters the control optimization step of GPS, which locally fits and exploits surrogate models of the dynamics, and employs the existing supervised learning step. The proposed solution incorporates new processes to ensure effective local modeling despite the disorganized nature of sample data in rough terrain locomotion. Demonstrations in simulation reveal that the resulting controller sustains the highly adaptive behavior necessary to reliably traverse rough terrain.

[1]  Kostas E. Bekris,et al.  Any-Axis Tensegrity Rolling via Symmetry-Reduced Reinforcement Learning , 2018, ISER.

[2]  Josep M. Porta,et al.  Path planning for active tensegrity structures , 2016 .

[3]  Brian T. Mirletz,et al.  Goal-Directed CPG-Based Control for Tensegrity Spines with Many Degrees of Freedom Traversing Irregular Terrain , 2015 .

[4]  Kostas E. Bekris,et al.  From Quasi-static to Kinodynamic Planning for Spherical Tensegrity Locomotion , 2017, ISRR.

[5]  Yuval Tassa,et al.  Synthesis and stabilization of complex behaviors through online trajectory optimization , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[6]  Shinichi Hirai,et al.  Crawling by body deformation of tensegrity structure robots , 2009, 2009 IEEE International Conference on Robotics and Automation.

[7]  Jeffrey M. Friesen,et al.  Steerable Locomotion Controller for Six-strut Icosahedral Tensegrity Robots , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Sergey Levine,et al.  Deep reinforcement learning for tensegrity robot locomotion , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[9]  Atil Iscen,et al.  Flop and roll: Learning robust goal-directed locomotion for a Tensegrity Robot , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  Kostas E. Bekris,et al.  INTEGRATING SIMULATED TENSEGRITY MODELS WITH EFFICIENT MOTION PLANNING FOR PLANETARY NAVIGATION , 2016 .

[11]  Alice M. Agogino,et al.  Inclined surface locomotion strategies for spherical tensegrity robots , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[12]  Hod Lipson,et al.  Morphological communication: exploiting coupled dynamics in a complex mechanical structure to achieve locomotion , 2009, Journal of The Royal Society Interface.

[13]  Suiping Zhou,et al.  An efficient locomotion strategy for six-strut tensegrity robots , 2017, 2017 13th IEEE International Conference on Control & Automation (ICCA).

[14]  Alice M. Agogino,et al.  Tensegrity Robot Locomotion Under Limited Sensory Inputs via Deep Reinforcement Learning , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[15]  Chandana Paul,et al.  Design and control of tensegrity robots for locomotion , 2006, IEEE Transactions on Robotics.

[16]  Kostas E. Bekris,et al.  Symmetric Reduction of Tensegrity Rover Dynamics for Efficient Data-Driven Control , 2018, Earth and Space 2018.

[17]  Jessica K. Hodgins,et al.  Hierarchical Aligned Cluster Analysis for Temporal Clustering of Human Motion , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Josep M. Mirats-Tur,et al.  Tensegrity frameworks: dynamic analysis review and open problems , 2009 .

[19]  Alice M. Agogino,et al.  Robust learning of tensegrity robot control for locomotion through form-finding , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Sergey Levine,et al.  Reset-free guided policy search: Efficient deep reinforcement learning with stochastic initial states , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[21]  Jeffrey M. Friesen,et al.  Design of SUPERball v2, a Compliant Tensegrity Robot for Absorbing Large Impacts , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[22]  Atil Iscen,et al.  Learning Tensegrity Locomotion Using Open-Loop Control Signals and Coevolutionary Algorithms , 2015, Artificial Life.

[23]  Sergey Levine,et al.  Learning Neural Network Policies with Guided Policy Search under Unknown Dynamics , 2014, NIPS.

[24]  Kostas E. Bekris,et al.  Discovering a Library of Rhythmic Gaits for Spherical Tensegrity Locomotion , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).