iGibson 2.0: Object-Centric Simulation for Robot Learning of Everyday Household Tasks

Recent research in embodied AI has been boosted by the use of simula1 tion environments to develop and train robot learning approaches. However, the 2 use of simulation has skewed the attention to tasks that only require what robotics 3 simulators can simulate: motion and physical contact. We present iGibson 2.0, an 4 open-source simulation environment that supports the simulation of a more diverse 5 set of household tasks through three key innovations. First, iGibson 2.0 supports 6 object states, including temperature, wetness level, cleanliness level, and toggled 7 and sliced states, necessary to cover a wider range of tasks. Second, iGibson 2.0 8 implements a set of predicate logic functions that map the simulator states to logic 9 states like Cooked or Soaked. Additionally, given a logic state, iGibson 2.0 can 10 sample valid physical states that satisfy it. This functionality can generate poten11 tially infinite instances of tasks with minimal effort from the users. The sampling 12 mechanism allows our scenes to be more densely populated with small objects in 13 semantically meaningful locations. Third, iGibson 2.0 includes a virtual reality 14 (VR) interface to immerse humans in its scenes to collect demonstrations. As a 15 result, we can collect demonstrations from humans on these new types of tasks, and 16 use them for imitation learning. We evaluate the new capabilities of iGibson 2.0 to 17 enable robot learning of novel tasks, in the hope of demonstrating the potential of 18 this new simulator to support new research in embodied AI. iGibson 2.0 and its 19 new dataset will be publicly available after the anonymous reviewing process. 20

[1]  Sergey Levine,et al.  Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning , 2019, CoRL.

[2]  Christian Duriez,et al.  SOFA: A Multi-Model Framework for Interactive Physical Simulation , 2012 .

[3]  Yohan Payan,et al.  SOFT TISSUE BIOMECHANICAL MODELING FOR COMPUTER ASSISTED SURGERY: CHALLENGES AND PERSPECTIVES , 2016 .

[4]  Jitendra Malik,et al.  Habitat: A Platform for Embodied AI Research , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[5]  Jérémie Dequidt,et al.  Software toolkit for modeling, simulation, and control of soft robots , 2017, Adv. Robotics.

[6]  David Held,et al.  SoftGym: Benchmarking Deep Reinforcement Learning for Deformable Object Manipulation , 2020, CoRL.

[7]  Sanja Fidler,et al.  VirtualHome: Simulating Household Activities Via Programs , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[8]  Craig A. Knoblock,et al.  PDDL-the planning domain definition language , 1998 .

[9]  Joseph DelPreto,et al.  Helping Robots Learn: A Human-Robot Master-Apprentice Model Using Demonstrations via Virtual Reality Teleoperation , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[10]  Ken Goldberg,et al.  Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation , 2017, ICRA.

[11]  Minghao Gou,et al.  OCRTOC: A Cloud-Based Competition and Benchmark for Robotic Grasping and Manipulation , 2021, IEEE Robotics and Automation Letters.

[12]  Song-Chun Zhu,et al.  VRKitchen: an Interactive 3D Virtual Environment for Task-oriented Learning , 2019, ArXiv.

[13]  Dean Pomerleau,et al.  ALVINN, an autonomous land vehicle in a neural network , 2015 .

[14]  Sergey Levine,et al.  Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor , 2018, ICML.

[15]  Alberto Garcia-Garcia,et al.  UnrealROX: an extremely photorealistic virtual reality environment for robotics simulations and synthetic data generation , 2018, Virtual Reality.

[16]  B. Tversky,et al.  Journal of Experimental Psychology : General VOL . 113 , No . 2 JUNE 1984 Objects , Parts , and Categories , 2005 .

[17]  Roozbeh Mottaghi,et al.  Rearrangement: A Challenge for Embodied AI , 2020, ArXiv.

[18]  Chuang Gan,et al.  ThreeDWorld: A Platform for Interactive Multi-Modal Physical Simulation , 2020, ArXiv.

[19]  Jitendra Malik,et al.  Gibson Env: Real-World Perception for Embodied Agents , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[20]  Andrew J. Davison,et al.  RLBench: The Robot Learning Benchmark & Learning Environment , 2019, IEEE Robotics and Automation Letters.

[21]  José García Rodríguez,et al.  The RobotriX: An Extremely Photorealistic and Very-Large-Scale Indoor Dataset of Sequences with Robot Trajectories and Interactions , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[22]  Yuval Tassa,et al.  MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[23]  Silvio Savarese,et al.  iGibson 1.0: A Simulation Environment for Interactive Tasks in Large Realistic Scenes , 2020, 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[24]  Leonidas J. Guibas,et al.  SAPIEN: A SimulAted Part-Based Interactive ENvironment , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[25]  Karin Schwab,et al.  Unity Game Development Essentials , 2016 .

[26]  L. Barsalou,et al.  Whither structured representation? , 1999, Behavioral and Brain Sciences.

[27]  Soha Hassoun,et al.  Evolving soft robotic locomotion in PhysX , 2009, GECCO '09.

[28]  Geoffrey J. Gordon,et al.  A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.

[29]  Richard Fikes,et al.  STRIPS: A New Approach to the Application of Theorem Proving to Problem Solving , 1971, IJCAI.

[30]  Roozbeh Mottaghi,et al.  ManipulaTHOR: A Framework for Visual Object Manipulation , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Roozbeh Mottaghi,et al.  Visual Room Rearrangement , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  Ali Farhadi,et al.  AI2-THOR: An Interactive 3D Environment for Visual AI , 2017, ArXiv.

[33]  Joshua B. Tenenbaum,et al.  The ThreeDWorld Transport Challenge: A Visually Guided Task-and-Motion Planning Benchmark Towards Physically Realistic Embodied AI , 2021, 2022 International Conference on Robotics and Automation (ICRA).

[34]  Roberto Mart'in-Mart'in,et al.  robosuite: A Modular Simulation Framework and Benchmark for Robot Learning , 2020, ArXiv.

[35]  Silvio Savarese,et al.  Interactive Gibson Benchmark: A Benchmark for Interactive Navigation in Cluttered Environments , 2020, IEEE Robotics and Automation Letters.