Enhanced feed-forward for a user aware multi-touch device

Common multi-touch devices guide the user with feedback visualization during or after a registered interaction. Feed-forward techniques are less frequently used or not common at all. Our approach aims at a continuous process in which the system is aware of the users before, during, and after an explicit interaction takes place. This opens up the possibility for novel scenarios of user centered applications. Our setup utilizes Microsofts's depth-camera Kinect to collect the user's posture data in combination with a multi-touch device. This is a low cost and easy to install approach for collecting detailed information about the people and their position in close proximity of a multi-touch table as well as the location of their physical contact. Based on this information, we propose five phases of interaction and analyze the sequence of input during a typical workflow. Eight application concepts show the relevance of these phases using appropriate forms of visualization and we evaluated three of those concepts in a user-study.

[1]  Eva Hornecker,et al.  “I don’t understand it either, but it is cool” - visitor interactions with a multi-touch table in a museum , 2008, 2008 3rd IEEE International Workshop on Horizontal Interactive Human Computer Systems.

[2]  Daniel J. Wigdor,et al.  Ripples: utilizing per-contact visualizations to improve user interaction with touch displays , 2009, UIST '09.

[3]  Masanori Sugimoto,et al.  RoboTable2: a novel programming environment using physical robots on a tabletop platform , 2011, Advances in Computer Entertainment Technology.

[4]  Kok-Lim Low,et al.  SandCanvas: a multi-touch art medium inspired by sand animation , 2011, CHI.

[5]  Dan Saffer Designing gestural interfaces , 2009 .

[6]  Xing-Dong Yang,et al.  TouchCuts and TouchZoom: enhanced target selection for touch displays using finger proximity sensing , 2011, CHI.

[7]  Marc Hassenzahl,et al.  How motivational orientation influences the evaluation and choice of hedonic and pragmatic interactive products: The role of regulatory focus , 2008, Interact. Comput..

[8]  Gudrun Klinker,et al.  KinectTouch: accuracy test for a very low-cost 2.5D multitouch tracking system , 2011, ITS '11.

[9]  Orit Shaer,et al.  Reality-based interaction: a framework for post-WIMP interfaces , 2008, CHI.

[10]  Maneesh Agrawala,et al.  FingerGlass: efficient multiscale interaction on multitouch screens , 2011, CHI.

[11]  Radu-Daniel Vatavu,et al.  Detecting and Tracking Multiple Users in the Proximity of Interactive Tabletops , 2008 .

[12]  Saul Greenberg,et al.  The Continuous Interaction Space: Interaction Techniques Unifying Touch and Gesture on and above a Digital Surface , 2011, INTERACT.

[13]  Marcus A. Magnor,et al.  Markerless Motion Capture using multiple Color-Depth Sensors , 2011, VMV.

[14]  Stephan Wensveen,et al.  But how, Donald, tell us how?: on the creation of meaning in interaction design through feedforward and inherent feedback , 2002, DIS '02.

[15]  Chris Harrison,et al.  OmniTouch: wearable multitouch interaction everywhere , 2011, UIST.

[16]  Yvonne Rogers,et al.  Enticing People to Interact with Large Public Displays in Public Spaces , 2003, INTERACT.

[17]  Darren Leigh,et al.  DiamondTouch: a multi-user touch technology , 2001, UIST '01.

[18]  Derrick J. Parkhurst,et al.  Enhancing Multi-user Interaction with Multi-touch Tabletop Displays Using Hand Tracking , 2008, First International Conference on Advances in Computer-Human Interaction.

[19]  Tovi Grossman,et al.  Medusa: a proximity-aware multi-touch tabletop , 2011, UIST.

[20]  Kathy Ryall,et al.  UbiTable: Impromptu Face-to-Face Collaboration on Horizontal Interactive Surfaces , 2003, UbiComp.

[21]  William Buxton,et al.  A three-state model of graphical input , 1990, INTERACT.

[22]  Hrvoje Benko,et al.  Combining multiple depth cameras and projectors for interactions on, above and between surfaces , 2010, UIST.

[23]  D. Kammer,et al.  Taxonomy and Overview of Multi-touch Frameworks: Architecture, Scope and Features , 2010 .

[24]  Rainer Groh,et al.  Liquid: Library for Interactive User Interface Development , 2011, Mensch & Computer.

[25]  Klaus H. Hinrichs,et al.  Poster: Interscopic multi-touch surfaces: Using bimanual interaction for intuitive manipulation of spatial data , 2009, 2009 IEEE Symposium on 3D User Interfaces.

[26]  Jeremy White Multi-touch Interfaces and Map Navigation , 2009 .

[27]  William Buxton,et al.  When it gets more difficult, use both hands: exploring bimanual curve manipulation , 2005, Graphics Interface.

[28]  Antti Oulasvirta,et al.  It's Mine, Don't Touch!: interactions at a large multi-touch display in a city centre , 2008, CHI.