3D selection with freehand gesture

The use of 3D computer graphics is important in a very wide range of applications. However, user interaction with 3D applications is still challenging and often does not lend itself to established techniques that have been developed primarily for 2D desktop interaction. Meanwhile, 3D user interfaces that rely on tracking hand-held devices or fiducial markers attached to the user are cumbersome or entirely inappropriate in some situations. These challenges may be addressed by refining and building on the increasing use of freehand gestural input, i.e. without markers or hand-held devices, to extend the fluidity and immediacy of today's 2D touch-based interactions. In this paper, we analyze the characteristics of freehand gestural 3D interaction, and report a set of 3 related evaluation studies focused on the fundamental user interface task of object selection. We found that interaction design requiring a high accuracy single action are not appropriate for freehand gestural selection, while separating it into several connected low demand operations could be a potential solution; that our Reach technique is accurate and potentially useful for option selection tasks with freehand gesture; and that strong directional effects influence performance and usability of both 2D and 3D option selection. We propose guidelines for designers of 3D freehand gestural interaction based on our evaluation results.

[1]  Terry Winograd,et al.  FlowMenu: combining command, text, and data entry , 2000, UIST '00.

[2]  Tovi Grossman,et al.  The design and evaluation of multitouch marking menus , 2010, CHI.

[3]  Gang Ren,et al.  3D Marking menu selection with freehand gestures , 2012, 2012 IEEE Symposium on 3D User Interfaces (3DUI).

[4]  George W. Fitzmaurice,et al.  PieCursor: merging pointing and command selection for rapid in-place tool switching , 2008, CHI.

[5]  Hrvoje Benko,et al.  Using Depth-Sensing Camera to Enable Freehand Interactions On and Above the Interactive Surface , 2008 .

[6]  François Guimbretière,et al.  Bimanual marking menu for near surface interactions , 2012, CHI.

[7]  Karin Coninx,et al.  Exploring the Effects of Environment Density and Target Visibility on Object Selection in 3D Virtual Environments , 2007, 2007 IEEE Symposium on 3D User Interfaces.

[8]  Ravin Balakrishnan,et al.  Simple vs. compound mark hierarchical marking menus , 2004, UIST '04.

[9]  Paul Richard,et al.  Evaluation of the Command and Control Cube , 2002, Proceedings. Fourth IEEE International Conference on Multimodal Interfaces.

[10]  Sriram Subramanian,et al.  GesText: accelerometer-based gestural text-entry systems , 2010, CHI.

[11]  Peng Song,et al.  A handle bar metaphor for virtual object manipulation with mid-air interaction , 2012, CHI.

[12]  D. Chaffin,et al.  An investigation of fitts' law using a wide range of movement amplitudes. , 1976, Journal of motor behavior.

[13]  Volker Dürr,et al.  Inter-joint coupling and joint angle synergies of human catching movements. , 2010, Human movement science.

[14]  Mark W. Newman,et al.  Escape: a target selection technique using visually-cued gestures , 2008, CHI.

[15]  William Buxton,et al.  The limits of expert performance using hierarchic marking menus , 1993, INTERCHI.

[16]  William Buxton,et al.  User learning and performance with marking menus , 1994, CHI 1994.

[17]  Patrick Baudisch,et al.  Precise selection techniques for multi-touch screens , 2006, CHI.

[18]  Shumin Zhai,et al.  More than dotting the i's --- foundations for crossing-based interfaces , 2002, CHI.

[19]  Hongan Wang,et al.  Tilt menu: using the 3D orientation information of pen devices to extend the selection capability of pen-based user interfaces , 2008, CHI.

[20]  Mark Billinghurst,et al.  Seamless interaction in space , 2011, OZCHI.

[21]  Andrew D. Wilson Using a depth camera as a touch sensor , 2010, ITS '10.

[22]  Andrew Wilson,et al.  MirageTable: freehand interaction on a projected augmented reality tabletop , 2012, CHI.

[23]  Abigail Sellen,et al.  An Empirical Evaluation of Some Articulatory and Cognitive Aspects of Marking Menus , 1993, Hum. Comput. Interact..

[24]  Chris North,et al.  AirStroke: bringing unistroke text entry to freehand gesture interfaces , 2011, CHI.

[25]  Doug A. Bowman,et al.  Rapid and accurate 3D selection by progressive refinement , 2011, 2011 IEEE Symposium on 3D User Interfaces (3DUI).

[26]  Brad A. Myers,et al.  EdgeWrite: a stylus-based text entry method designed for high accuracy and stability of motion , 2003, UIST '03.

[27]  Sungjae Hwang,et al.  Aroundplot: Focus+context interface for off-screen objects in 3D environments , 2011, Comput. Graph..

[28]  Maneesh Agrawala,et al.  Zone and polygon menus: using relative position to increase the breadth of multi-stroke marking menus , 2006, CHI.

[29]  Hrvoje Benko,et al.  Multi-point interactions with immersive omnidirectional visualizations in a dome , 2010, ITS '10.

[30]  Tovi Grossman,et al.  Multi-finger gestural interaction with 3d volumetric displays , 2004, UIST '04.

[31]  Sriram Subramanian,et al.  Tilt techniques: investigating the dexterity of wrist-based input , 2009, CHI.

[32]  Doug A. Bowman,et al.  Design and evaluation of menu systems for immersive virtual environments , 2001, Proceedings IEEE Virtual Reality 2001.

[33]  M. Sheelagh T. Carpendale,et al.  Gestures in the wild: studying multi-touch gesture sequences on interactive tabletop exhibits , 2011, CHI.

[34]  Joseph J. LaViola,et al.  An exploration of menu techniques using a 3D game input device , 2009, FDG.

[35]  Mathias Kölsch,et al.  The Postural Comfort Zone for Reaching Gestures , 2003 .

[36]  Jacob O. Wobbrock,et al.  Enhanced area cursors: reducing fine pointing demands for people with motor impairments , 2010, UIST.

[37]  Chris North,et al.  Design and evaluation of freehand menu selection interfaces using tilt and pinch gestures , 2011, Int. J. Hum. Comput. Stud..

[38]  Sriram Subramanian,et al.  Talking about tactile experiences , 2013, CHI.

[39]  Andrew Bragdon,et al.  Gesture select:: acquiring remote targets on large displays without pointing , 2011, CHI.

[40]  Tovi Grossman,et al.  The design and evaluation of selection techniques for 3D volumetric displays , 2006, UIST.

[41]  Tovi Grossman,et al.  Exploring the design space of multiscale 3D orientation , 2010, AVI.

[42]  Donald A. Norman,et al.  Natural user interfaces are not natural , 2010, INTR.

[43]  Kellogg S. Booth,et al.  Mid-air text input techniques for very large wall displays , 2009, Graphics Interface.

[44]  Hrvoje Benko,et al.  Beyond flat surface computing: challenges of depth-aware and curved interfaces , 2009, ACM Multimedia.

[45]  Daniel Vogel,et al.  Distant freehand pointing and clicking on very large, high resolution displays , 2005, UIST.

[46]  J. Edward Swan,et al.  Toward Disambiguating Multiple Selections for Frustum-Based Pointing , 2006, 3D User Interfaces (3DUI'06).

[47]  M. Weiser,et al.  An empirical comparison of pie vs. linear menus , 1988, CHI '88.

[48]  Chunxin Yang,et al.  An algorithm for simulating human arm movement considering the comfort level , 2005, Simul. Model. Pract. Theory.

[49]  Kaushik Das,et al.  An evaluation of menu properties and pointing techniques in a projection-based VR environment , 2010, 2010 IEEE Symposium on 3D User Interfaces (3DUI).

[50]  Matthias Bues,et al.  iSith - Intersection-based Spatial Interaction for Two Hands , 2006, 3D User Interfaces (3DUI'06).

[51]  Daniel Vogel,et al.  Interactive public ambient displays: transitioning from implicit to explicit, public to personal, interaction with multiple users , 2004, UIST '04.

[52]  Chris Harrison,et al.  OmniTouch: wearable multitouch interaction everywhere , 2011, UIST.

[53]  Robert J. Teather,et al.  Assessing the Effects of Orientation and Device on (Constrained) 3D Movement Techniques , 2008, 2008 IEEE Symposium on 3D User Interfaces.

[54]  Tovi Grossman,et al.  Pointing at trivariate targets in 3D environments , 2004, CHI.

[55]  Raimund Dachselt,et al.  Three-dimensional menus: A survey and taxonomy , 2007, Comput. Graph..

[56]  Andrew D. Wilson Robust computer vision-based detection of pinching for one and two-handed gesture input , 2006, UIST.

[57]  Mark Green,et al.  JDCAD: A highly interactive 3D modeling system , 1994, Comput. Graph..

[58]  Krzysztof Z. Gajos,et al.  Goal Crossing with Mice and Trackballs for People with Motor Impairments: Performance, Submovements, and Design Directions , 2008, TACC.

[59]  Jock D. Mackinlay,et al.  A morphological analysis of the design space of input devices , 1991, TOIS.