Full-Body Visual Self-Modeling of Robot Morphologies

Internal computational models of physical bodies are fundamental to the ability of robots and animals alike to plan and control their actions. These “selfmodels” allow robots to consider outcomes of multiple possible future actions, without trying them out in physical reality. Recent progress in fully datadriven self-modeling has enabled machines to learn their own forward kinematics directly from task-agnostic interaction data. However, forward-kinematics models can only predict limited aspects of the morphology, such as the position of end effectors or velocity of joints and masses. A key challenge is to model the entire morphology and kinematics, without prior knowledge of what aspects of the morphology will be relevant to future tasks. Here, we propose that instead of directly modeling forward-kinematics, a more useful form of self-modeling is one that could answer space occupancy queries, conditioned on the robot’s state. Such query-driven self models are continuous in the spatial domain, memory efficient, fully differentiable and kinematic aware. In physical experiments, we demonstrate how a visual self-model is accurate to about one percent of the workspace, enabling the robot to perform various motion planning and control tasks. Visual self-modeling can also allow the robot to detect, localize and recover from real-world damage, leading to improved machine resiliency.

[1]  Kaiyu Hang,et al.  Manipulation for self-Identification, and self-Identification for better manipulation , 2021, Science Robotics.

[2]  Hod Lipson,et al.  Smile Like You Mean It: Driving Animatronic Robotic Face with Learned Models , 2021, 2021 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Jan-Michael Frahm,et al.  Structure-from-Motion Revisited , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Christian Duriez,et al.  SOFA: A Multi-Model Framework for Interactive Physical Simulation , 2012 .

[5]  Raia Hadsell,et al.  Graph networks as learnable physics engines for inference and control , 2018, ICML.

[6]  Yohan Payan,et al.  SOFT TISSUE BIOMECHANICAL MODELING FOR COMPUTER ASSISTED SURGERY: CHALLENGES AND PERSPECTIVES , 2016 .

[7]  Hod Lipson,et al.  Resilient Machines Through Continuous Self-Modeling , 2006, Science.

[8]  Ruzena Bajcsy,et al.  Active touch and robot perception , 1984 .

[9]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[10]  Gordon Wetzstein,et al.  Implicit Neural Representations with Periodic Activation Functions , 2020, NeurIPS.

[11]  Hod Lipson,et al.  Automatic Synthesis of Multiple Internal Models Through Active Exploration , 2005 .

[12]  Jan-Michael Frahm,et al.  Pixelwise View Selection for Unstructured Multi-View Stereo , 2016, ECCV.

[13]  Emilio Frazzoli,et al.  Incremental Sampling-based Algorithms for Optimal Motion Planning , 2010, Robotics: Science and Systems.

[14]  G. G. Gallup,et al.  Self‐awareness and the emergence of mind in primates , 1982, American journal of primatology.

[15]  Richard A. Newcombe,et al.  DeepSDF: Learning Continuous Signed Distance Functions for Shape Representation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  Andrew Howard,et al.  Design and use paradigms for Gazebo, an open-source multi-robot simulator , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[17]  Hod Lipson,et al.  Task-agnostic self-modeling machines , 2019, Science Robotics.

[18]  Santhosh K. Ramakrishnan,et al.  Emergence of exploratory look-around behaviors through active observation completion , 2019, Science Robotics.

[19]  Jonathan T. Barron,et al.  Learned Initializations for Optimizing Coordinate-Based Neural Representations , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  William E. Lorensen,et al.  Marching cubes: A high resolution 3D surface construction algorithm , 1987, SIGGRAPH.

[21]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[22]  Yuval Tassa,et al.  MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[23]  P. Rochat Five levels of self-awareness as they unfold early in life , 2003, Consciousness and Cognition.

[24]  Oliver Brock,et al.  Interactive Perception: Leveraging Action in Perception and Perception in Action , 2016, IEEE Transactions on Robotics.

[25]  Siddhartha S. Srinivasa,et al.  DART: Dynamic Animation and Robotics Toolkit , 2018, J. Open Source Softw..

[26]  H. Gardner,et al.  Frames of Mind: The Theory of Multiple Intelligences , 1983 .