Inferring Generalized Pick-and-Place Tasks from Pointing Gestures

In this paper, we describe how to simplify the instruction of robots to perform pick-and-place tasks. This is achieved by inferring additional information about the environment, locations in it, and the used object, based on semantic knowledge. We use the Kinect RGB-D camera mounted on a PR2 robot and the human tracking system OpenNI that allows the interpretation of simple human gestures. Our application builds enables virtual movements of objects freely in space and the use of an ontology for reasoning about the plausibility of the tasks specified by releasing objects in certain locations. Our experiments show how relatively complex object manipulation tasks can be specified as simply as “drag-and-drop”.

[1]  Moritz Tenorth,et al.  KNOWROB-MAP - knowledge-linked semantic object maps , 2010, 2010 10th IEEE-RAS International Conference on Humanoid Robots.

[2]  Matei T. Ciocarlie,et al.  Towards Reliable Grasping and Manipulation in Household Environments , 2010, ISER.

[3]  Moritz Tenorth,et al.  CRAM — A Cognitive Robot Abstract Machine for everyday manipulation in human environments , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[5]  Nico Blodow,et al.  Towards 3D Point cloud based object maps for household environments , 2008, Robotics Auton. Syst..

[6]  Moritz Tenorth,et al.  KNOWROB — knowledge processing for autonomous personal robots , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Dejan Pangercic,et al.  Combining perception and knowledge processing for everyday manipulation , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[8]  Advait Jain,et al.  A clickable world: Behavior selection through pointing and context for mobile manipulation , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[9]  Nico Blodow,et al.  Model-based and learned semantic object labeling in 3D point cloud maps of kitchen environments , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  David Nistér,et al.  Scalable Recognition with a Vocabulary Tree , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[11]  Joachim Hertzberg,et al.  Towards semantic maps for mobile robots , 2008, Robotics Auton. Syst..

[12]  Michael Beetz,et al.  Real-time perception-guided motion planning for a personal robot , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Yasuo Kuniyoshi,et al.  AI Goggles: Real-time Description and Retrieval in the Real World with Online Learning , 2009, 2009 Canadian Conference on Computer and Robot Vision.

[14]  S. Srinivasa,et al.  Push-grasping with dexterous hands: Mechanics and a method , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.