Leveraging depth data in remote robot teleoperation interfaces for general object manipulation

Robust remote teleoperation of high-degree-of-freedom manipulators is of critical importance across a wide range of robotics applications. Contemporary robot manipulation interfaces primarily utilize a free positioning pose specification approach to independently control each degree of freedom in free space. In this work, we present two novel interfaces, constrained positioning and point-and-click. Both novel approaches incorporate scene information from depth data into the grasp pose specification process, effectively reducing the number of 3D transformations the user must input. The novel interactions are designed for 2D image streams, rather than traditional 3D virtual scenes, further reducing mental transformations by eliminating the controllable camera viewpoint in favor of fixed physical camera viewpoints. We present interface implementations of our novel approaches, as well as free positioning, in both 2D and 3D visualization modes. In addition, we present results of a 90-participant user study evaluation comparing the effectiveness of each approach for a set of general object manipulation tasks, and the effects of implementing each approach in 2D image views versus 3D depth views. The results of our study show that point-and-click outperforms both free positioning and constrained positioning by significantly increasing the number of tasks completed and significantly reducing task failures and grasping errors, while significantly reducing the number of user interactions required to specify poses. In addition, we found that regardless of the interaction approach, the 2D visualization mode resulted in significantly better performance than the 3D visualization mode, with statistically significant reductions in task failures, grasping errors, task completion time, number of interactions, and user workload, all while reducing bandwidth requirements imposed by streaming depth data.

[1]  Carlo Alberto Avizzano,et al.  Augmented reality-aided tele-presence system for robot manipulation in industrial manufacturing , 2015, VRST.

[2]  Matei T. Ciocarlie,et al.  Mobile manipulation through an assistive home robot , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[3]  Markus Vincze,et al.  Learning grasps with topographic features , 2015, Int. J. Robotics Res..

[4]  Robert Platt,et al.  Using Geometry to Detect Grasp Poses in 3D Point Clouds , 2015, ISRR.

[5]  Martin Hachet,et al.  Navidget for Easy 3D Camera Positioning from 2D Inputs , 2008, 2008 IEEE Symposium on 3D User Interfaces.

[6]  Sonia Chernova,et al.  Construction of a 3D object recognition and manipulation database from grasp demonstrations , 2016, Auton. Robots.

[7]  Daniela Rus,et al.  Baxter's Homunculus: Virtual Reality Spaces for Teleoperation in Manufacturing , 2017, IEEE Robotics and Automation Letters.

[8]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[9]  Eric Huang,et al.  A General‐purpose System for Teleoperation of the DRC‐HUBO Humanoid Robot , 2015, J. Field Robotics.

[10]  Dmitry Berenson,et al.  Toward a user-guided manipulation framework for high-DOF robots with limited communication , 2013, 2013 IEEE Conference on Technologies for Practical Robot Applications (TePRA).

[11]  Masahiro Hori,et al.  Smartphone game controller (マルチメディア・仮想環境基礎) , 2011 .

[12]  Shumin Zhai,et al.  Quantifying coordination in multiple DOF movement and its application to evaluating 6 DOF input devices , 1998, CHI.

[13]  Bilge Mutlu,et al.  A Motion Retargeting Method for Effective Mimicry-Based Teleoperation of Robot Arms , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[14]  Honglak Lee,et al.  Deep learning for detecting robotic grasps , 2013, Int. J. Robotics Res..

[15]  S. Hart,et al.  Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research , 1988 .

[16]  Paul Milgram,et al.  Measuring the allocation of control in a 6 degree-of-freedom docking experiment , 2000, CHI.

[17]  Thomas B. Sheridan,et al.  Space teleoperation through time delay: review and prognosis , 1993, IEEE Trans. Robotics Autom..

[18]  Robert Platt,et al.  Localizing Handle-Like Grasp Affordances in 3D Point Clouds , 2014, ISER.

[19]  Sonia Chernova,et al.  Robot Web Tools: Efficient messaging for cloud robotics , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[20]  Christopher G. Atkeson,et al.  Human‐in‐the‐loop Control of a Humanoid Robot for Disaster Response: A Report from the DARPA Robotics Challenge Trials , 2015, J. Field Robotics.

[21]  J. Edward Colgate,et al.  Improving teleoperation: reducing mental rotations and translations , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[22]  Xinyu Liu,et al.  Dex-Net 2.0: Deep Learning to Plan Robust Grasps with Synthetic Point Clouds and Analytic Grasp Metrics , 2017, Robotics: Science and Systems.

[23]  Charles C. Kemp,et al.  Assistive Mobile Manipulation : Designing for Operators with Motor Impairments Phillip , 2016 .

[24]  Lawson L. S. Wong,et al.  Learning Grasp Strategies with Partial Shape Information , 2008, AAAI.

[25]  Holly A. Yanco,et al.  Analysis of Human‐robot Interaction at the DARPA Robotics Challenge Trials , 2015, J. Field Robotics.

[26]  Jean Scholtz,et al.  Analysis of human–robot interaction at the DARPA Robotics Challenge Finals , 2017, Int. J. Robotics Res..

[27]  Abhinav Gupta,et al.  Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[28]  Doug A. Bowman,et al.  Human‐robot Teaming for Rescue Missions: Team ViGIR's Approach to the 2013 DARPA Robotics Challenge Trials , 2015, J. Field Robotics.

[29]  Sergey Levine,et al.  Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection , 2016, Int. J. Robotics Res..

[30]  Matei T. Ciocarlie,et al.  Contact-reactive grasping of objects with partial shape information , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[31]  Gianluca Paravati,et al.  Mixed Reality-Based User Interaction Feedback for a Hand-Controlled Interface Targeted to Robot Teleoperation , 2017, AVR.

[32]  L. Parsons Inability to Reason About an Object's Orientation Using an Axis and Angle of Rotation , 1995 .

[33]  Leila Takayama,et al.  Strategies for human-in-the-loop robotic grasping , 2012, 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[34]  Ashutosh Saxena,et al.  Efficient grasping from RGBD images: Learning using a new rectangle representation , 2011, 2011 IEEE International Conference on Robotics and Automation.

[35]  Matei T. Ciocarlie,et al.  Interactive Markers: 3-D User Interfaces for ROS Applications [ROS Topics] , 2011, IEEE Robotics Autom. Mag..

[36]  Mark Billinghurst,et al.  Using a mobile phone for 6 DOF mesh editing , 2007, CHINZ.

[37]  Sachin Chitta,et al.  MoveIt!: An Introduction , 2016 .

[38]  Tiffany L. Chen,et al.  Robots for Humanity : A Case Study in Assistive Mobile Manipulation , 2012 .