Qualitative Spatial Ontologies for Robot Dynamics

Following Smith and Gasser’s work on embodied cognition, one can consider a robot as an intelligent agent that interacts with its external environment through sensorimotor activities, such as touching, lifting, standing, sitting, and walking. In this paper we explore the ontologies that are required to represent and reason about robot dynamics. We propose new ontologies for robotic components and poses, including a new nonclassical mereotopology for touch contact. The design of the ontologies is driven by semantic parsing of natural language instructions (e.g. “Lift the box that is beside the chair and place it on the table"), through which we identify the spatial and mereotopological relations among a robot’s components and the external world, as well as the activities that the robot can perform to change these relationships.

[1]  Roberto Casati,et al.  Parts and Places: The Structures of Spatial Representation , 1999 .

[2]  Jonathan Berant,et al.  Building a Semantic Parser Overnight , 2015, ACL.

[3]  P. Alam,et al.  H , 1887, High Explosives, Propellants, Pyrotechnics.

[4]  Luke S. Zettlemoyer,et al.  Learning to Parse Natural Language Commands to a Robot Control System , 2012, ISER.

[5]  Peter Stone,et al.  Learning to Interpret Natural Language Commands through Human-Robot Dialog , 2015, IJCAI.

[6]  Michael Grüninger,et al.  Robot Meets World , 2020, JOWO.

[7]  장윤희,et al.  Y. , 2003, Industrial and Labor Relations Terms.

[8]  P. Alam ‘G’ , 2021, Composites Engineering: An A–Z Guide.

[9]  Michael Gasser,et al.  The Development of Embodied Cognition: Six Lessons from Babies , 2005, Artificial Life.

[10]  Roozbeh Mottaghi,et al.  ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  Stefan Lee,et al.  Embodied Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[12]  P. Alam ‘S’ , 2021, Composites Engineering: An A–Z Guide.

[13]  Qi Wu,et al.  Vision-and-Language Navigation: Interpreting Visually-Grounded Navigation Instructions in Real Environments , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[14]  Danna Zhou,et al.  d. , 1840, Microbial pathogenesis.

[15]  Laure Vieu,et al.  Toward a Geometry of Common Sense: A Semantics and a Complete Axiomatization of Mereotopology , 1995, IJCAI.

[16]  Matthew R. Walter,et al.  Understanding Natural Language Commands for Robotic Navigation and Mobile Manipulation , 2011, AAAI.

[17]  Andrew Chou,et al.  Semantic Parsing on Freebase from Question-Answer Pairs , 2013, EMNLP.

[18]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[19]  Yuandong Tian,et al.  Building Generalizable Agents with a Realistic and Rich 3D Environment , 2018, ICLR.

[20]  Jitendra Malik,et al.  Habitat: A Platform for Embodied AI Research , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[21]  Matthew R. Walter,et al.  Learning Semantic Maps from Natural Language Descriptions , 2013, Robotics: Science and Systems.