Extraction of Physically Plausible Support Relations to Predict and Validate Manipulation Action Effects

Reliable execution of robot manipulation actions in cluttered environments requires that the robot is able to understand relations between objects and reason about consequences of actions applied to these objects. We present an approach for extracting physically plausible support relations between objects based on visual information, which does not require any prior knowledge about physical object properties, e.g., mass distribution or friction coefficients. Based on a scene representation enriched by such physically plausible support relations between objects, we derive predictions about action effects. These predictions take into account uncertainty about support relations and allow applying strategies for safe bimanual object manipulation when needed. The extraction of physically plausible support relations is evaluated both in simulation and in real-world experiments using real data from a depth camera, whereas the handling of support relation uncertainties is validated on the humanoid robot ARMAR-III.

[1]  Patrick J. Hayes,et al.  The Naive Physics Manifesto , 1990, The Philosophy of Artificial Intelligence.

[2]  S. Vosniadou On the Nature of Naïve Physics , 2002 .

[3]  Tamim Asfour,et al.  ARMAR-III: An Integrated Humanoid Platform for Sensory-Motor Control , 2006, 2006 6th IEEE-RAS International Conference on Humanoid Robots.

[4]  Mark Steedman,et al.  Using Kernel Perceptrons to Learn Action Effects for Planning , 2008 .

[5]  Carl Barck-Holst,et al.  Fitting Primitive Shapes to Point Clouds for Robotic Grasping , 2009 .

[6]  Florentin Wörgötter,et al.  Cognitive agents - a procedural perspective relying on the predictability of Object-Action-Complexes (OACs) , 2009, Robotics Auton. Syst..

[7]  Benjamin Rosman,et al.  Learning spatial relationships between objects , 2011, Int. J. Robotics Res..

[8]  Mark Steedman,et al.  Object-Action Complexes: Grounded abstractions of sensory-motor processes , 2011, Robotics Auton. Syst..

[9]  Radu Bogdan Rusu,et al.  3D is here: Point Cloud Library (PCL) , 2011, 2011 IEEE International Conference on Robotics and Automation.

[10]  Derek Hoiem,et al.  Indoor Segmentation and Support Inference from RGBD Images , 2012, ECCV.

[11]  C. V. Jawahar,et al.  Learning support order for manipulation in clutter , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  Katsushi Ikeuchi,et al.  Beyond Point Clouds: Scene Understanding by Reasoning Geometry and Physics , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  Katsushi Ikeuchi,et al.  Scene Understanding by Reasoning Stability and Safety , 2015, International Journal of Computer Vision.

[14]  Mirko Wächter,et al.  The robot software framework ArmarX , 2015, it Inf. Technol..

[15]  Justus H. Piater,et al.  Bottom-up learning of object categories, action effects and logical rules: From continuous manipulative exploration to symbolic planning , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Tsuhan Chen,et al.  3D Reasoning from Blocks to Stability , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  Erik Schaffernicht,et al.  Support relation analysis and decision making for safe robotic manipulation tasks , 2015, Robotics Auton. Syst..

[18]  Andreas Birk,et al.  Physics-based damage-aware manipulation strategy planning using Scene Dynamics Anticipation , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[19]  C. V. Jawahar,et al.  Single and Multiple View Support Order Prediction in Clutter for Manipulation , 2016, J. Intell. Robotic Syst..

[20]  Jitendra Malik,et al.  Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.

[21]  Dieter Fox,et al.  SE3-nets: Learning rigid body motion using deep neural networks , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[22]  Eren Erdal Aksoy,et al.  Graph-based visual semantic perception for humanoid robots , 2017, 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids).

[23]  Wolfram Burgard,et al.  Learning to Singulate Objects using a Push Proposal Network , 2017, ISRR.