Natural Language Based Multimodal Interface for UAV Mission Planning

As the number of viable applications for unmanned aerial vehicle (UAV) systems increases at an exponential rate, interfaces that reduce the reliance on highly skilled engineers and pilots must be developed. Recent work aims to make use of common human communication modalities such as speech and gesture. This paper explores a multimodal natural language interface that uses a combination of speech and gesture input modalities to build complex UAV flight paths by defining trajectory segment primitives. Gesture inputs are used to define the general shape of a segment while speech inputs provide additional geometric information needed to fully characterize a trajectory segment. A user study is conducted in order to evaluate the efficacy of the multimodal interface.

[1]  S. Hart,et al.  Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research , 1988 .

[2]  Kenji Shimada,et al.  A Natural Interaction Interface for UAVs Using Intuitive Gesture Recognition , 2017 .

[3]  Alin Albu-Schäffer,et al.  A human-centered approach to robot gesture based communication within collaborative working processes , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  Yael Edan,et al.  Vision-based hand-gesture applications , 2011, Commun. ACM.

[5]  Ehud Sharlin,et al.  Flying Frustum: A Spatial Interface for Enhancing Human-UAV Awareness , 2015, HAI.

[6]  Martin Molina,et al.  Natural user interfaces for human-drone multi-modal interaction , 2016, 2016 International Conference on Unmanned Aircraft Systems (ICUAS).

[7]  Erica L. Meszaros,et al.  'Fly Like This': Natural Language Interface for UAV Mission Planning , 2017, ACHI 2017.

[8]  Randal W. Beard,et al.  Semi-autonomous human-UAV interfaces for fixed-wing mini-UAVs , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[9]  Daniel Cremers,et al.  FollowMe: Person following and gesture recognition with a quadrocopter , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  Magdalena D. Bugajska,et al.  Building a Multimodal Human-Robot Interface , 2001, IEEE Intell. Syst..

[11]  Yan Li,et al.  A Survey of Autonomous Control for UAV , 2009, 2009 International Conference on Artificial Intelligence and Computational Intelligence.

[12]  James A. Landay,et al.  Drone & me: an exploration into natural human-drone interaction , 2015, UbiComp.

[13]  Susan G. Hill,et al.  Traditional and raw task load index (TLX) correlations: Are paired comparisons necessary? In A , 1989 .

[14]  Wendell H. Chun,et al.  Team-Centered Virtual Interactive Presence for Adjustable Autonomy , 2005 .

[15]  Timothy Brittain-Catlin Put it there , 2013 .

[16]  Richard A. Bolt,et al.  “Put-that-there”: Voice and gesture at the graphics interface , 1980, SIGGRAPH '80.

[17]  Ehud Sharlin,et al.  Collocated interaction with flying robots , 2011, 2011 RO-MAN.