Human Gaze-Driven Spatial Tasking of an Autonomous MAV

In this letter, we address the problem of providing human-assisted quadrotor navigation using a set of eye tracking glasses. The advent of these devices (i.e., eye tracking glasses, virtual reality tools, etc.) provides the opportunity to create new, noninvasive forms of interaction between humans and robots. We show how a set of glasses equipped with gaze tracker, a camera, and an inertial measurement unit (IMU) can be used to estimate the relative position of the human with respect to a quadrotor, and decouple the gaze direction from the head orientation, which allows the human to spatially task (i.e., send new 3-D navigation waypoints to) the robot in an uninstrumented environment. We decouple the gaze direction from head motion by tracking the human's head orientation using a combination of camera and IMU data. In order to detect the flying robot, we train and use a deep neural network. We experimentally evaluate the proposed approach, and show that our pipeline has the potential to enable gaze-driven autonomy for spatial tasking. The proposed approach can be employed in multiple scenarios including inspection and first response, as well as by people with disabilities that affect their mobility.

[1]  Luca Maria Gambardella,et al.  Human Control of UAVs using Face Pose Estimates and Hand Gestures , 2014, 2014 9th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[2]  Joseph H. Goldberg,et al.  Identifying fixations and saccades in eye-tracking protocols , 2000, ETRA.

[3]  Morten Fjeld,et al.  Social Drone Companion for the Home Environment: a User-Centric Exploration , 2017, HAI.

[4]  Carlo Tomasi,et al.  Good features to track , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[5]  S. Brennan,et al.  Speakers' eye gaze disambiguates referring expressions early during face-to-face conversation , 2007 .

[6]  B. Scassellati,et al.  Social eye gaze in human-robot interaction , 2017, HRI 2017.

[7]  A. Young,et al.  Reading the mind from eye gaze , 2002, Neuropsychologia.

[8]  Marc Pollefeys,et al.  A Minimal Case Solution to the Calibrated Relative Pose Problem for the Case of Two Known Orientation Angles , 2010, ECCV.

[9]  Vijay Kumar,et al.  Autonomous Navigation and Mapping for Inspection of Penstocks and Tunnels With MAVs , 2017, IEEE Robotics and Automation Letters.

[10]  Richard M. Murray,et al.  A Mathematical Introduction to Robotic Manipulation , 1994 .

[11]  Tom Drummond,et al.  Machine Learning for High-Speed Corner Detection , 2006, ECCV.

[12]  Greg Mori,et al.  HRI in the sky: Creating and commanding teams of UAVs with a vision-mediated gestural interface , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Ehud Sharlin,et al.  Collocated interaction with flying robots , 2011, 2011 RO-MAN.

[14]  Peter Stone,et al.  Inferring User Intention using Gaze in Vehicles , 2018, ICMI.

[15]  P. Fletcher,et al.  Seeing other minds: attributed mental states influence perception , 2010, Trends in Cognitive Sciences.

[16]  Matthias Scheutz,et al.  Adaptive eye gaze patterns in interactions with human and artificial agents , 2012, TIIS.

[17]  Yu Wang,et al.  Human-Robot Interaction Based on Gaze Gestures for the Drone Teleoperation , 2014 .

[18]  Shaojie Shen,et al.  SLAM-based localization of 3D gaze using a mobile eye tracker , 2018, ETRA.

[19]  Luca Maria Gambardella,et al.  Human-swarm interaction using spatial gestures , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[20]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[21]  I. Scott MacKenzie,et al.  The use of gaze to control drones , 2014, ETRA.

[22]  N. Clayton,et al.  Mental-state attribution drives rapid, reflexive gaze following , 2010, Attention, perception & psychophysics.

[23]  Vijay Kumar,et al.  Cooperative localization and mapping of MAVs using RGB-D sensors , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[24]  Giuseppe Loianno,et al.  A Distributed Optimization Framework for Localization and Formation Control: Applications to Vision-Based Measurements , 2016, IEEE Control Systems.

[25]  Dieter Schmalstieg,et al.  Drone-Augmented Human Vision: Exocentric Control for Drones Exploring Hidden Areas , 2018, IEEE Transactions on Visualization and Computer Graphics.

[26]  Joseph J. LaViola,et al.  Exploring 3d gesture metaphors for interaction with unmanned aerial vehicles , 2013, IUI '13.

[27]  M. Crocker,et al.  Investigating joint attention mechanisms through spoken human–robot interaction , 2011, Cognition.

[28]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[29]  Kevin Lee,et al.  Come See This! Augmented Reality to Enable Human-Robot Cooperative Search , 2018, 2018 IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR).

[30]  Roland Siegwart,et al.  Extending kalibr: Calibrating the extrinsics of multiple IMUs and of individual axes , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[31]  Peter Stone,et al.  A Study of Human-Robot Copilot Systems for En-route Destination Changing , 2018, 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[32]  Richard T. Vaughan,et al.  Ready—Aim—Fly! Hands-Free Face-Based HRI for 3D Trajectory Control of UAVs , 2017, 2017 14th Conference on Computer and Robot Vision (CRV).

[33]  Darius Burschka,et al.  Toward a Fully Autonomous UAV: Research Platform for Indoor and Outdoor Urban Search and Rescue , 2012, IEEE Robotics & Automation Magazine.

[34]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[35]  Vijay Kumar,et al.  Estimation, Control, and Planning for Aggressive Flight With a Small Quadrotor With a Single Camera and IMU , 2017, IEEE Robotics and Automation Letters.

[36]  Jennifer Lee,et al.  Communicating Robot Motion Intent with Augmented Reality , 2018, 2018 13th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[37]  Edward G. Freedman,et al.  Coordination of the eyes and head: movement kinematics , 2000, Experimental Brain Research.

[38]  Giuseppe Loianno,et al.  Autonomous navigation of micro aerial vehicles using high-rate and low-cost sensors , 2018, Auton. Robots.

[39]  Kazuya Yoshida,et al.  Collaborative mapping of an earthquake‐damaged building via ground and aerial robots , 2012, J. Field Robotics.

[40]  Lorenzo Marconi,et al.  Impedance control of an aerial manipulator , 2012, 2012 American Control Conference (ACC).

[41]  Tony Belpaeme,et al.  Head Pose Estimation is an Inadequate Replacement for Eye Gaze in Child-Robot Interaction , 2015, HRI.