Development of actuated Tangible User Interfaces: new interaction concepts and evaluation methods

Making information understandable and literally graspable is the main goal of tangible interaction research. By giving digital data physical representations (Tangible User Interface Objects, or TUIOs), they can be used and manipulated like everyday objects with the users’ natural manipulation skills. Such physical interaction is basically of uni-directional kind, directed from the user to the system, limiting the possible interaction patterns. In other words, the system has no means to actively support the physical interaction. Within the frame of tabletop tangible user interfaces, this problem was addressed by the introduction of actuated TUIOs, that are controllable by the system. Within the frame of this thesis, we present the development of our own actuated TUIOs and address multiple interaction concepts we identified as research gaps in literature on actuated Tangible User Interfaces (TUIs). Gestural interaction is a natural means for humans to non-verbally communicate using their hands. TUIs should be able to support gestural interaction, since our hands are already heavily involved in the interaction. This has rarely been investigated in literature. For a tangible social network client application, we investigate two methods for collecting user-defined gestures that our system should be able to interpret for triggering actions. Versatile systems often understand a wide palette of commands. Another approach for triggering actions is the use of menus. We explore the design space of menu metaphors used in TUIs and present our own actuated dial-based approach. Rich interaction modalities may support the understandability of the represented data and make the interaction with them more appealing, but also mean high demands on real-time precessing. We highlight new research directions for integrated feature rich and multi-modal interaction, such as graphical display, sound output, tactile feedback, our actuated menu and automatically maintained relations between actuated TUIOs within a remote collaboration application. We also tackle the introduction of further sophisticated measures for the evaluation of TUIs to provide further evidence to the theories on tangible interaction. We tested our enhanced measures within a comparative study. Since one of the key factors in effective manual interaction is speed, we benchmarked both the human hand’s manipulation speed and compare it with the capabilities of our own implementation of actuated TUIOs and the systems described in literature. After briefly discussing applications that lie beyond the scope of this thesis, we conclude with a collection of design guidelines gathered in the course of this work and integrate them together with our findings into a larger frame.

[1]  Bill Buxton,et al.  Multi-Touch Systems that I Have Known and Loved , 2009 .

[2]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[3]  Lars Erik Holmquist,et al.  Token-Based Acces to Digital Information , 1999, HUC.

[4]  Heekuck Oh,et al.  Neural Networks for Pattern Recognition , 1993, Adv. Comput..

[5]  William Buxton,et al.  Graspable user interfaces , 1996 .

[6]  Anders Green,et al.  Social and collaborative aspects of interaction with a service robot , 2003, Robotics Auton. Syst..

[7]  Helge J. Ritter,et al.  Saving and Restoring Mechanisms for Tangible User Interfaces through Tangible Active Objects , 2011, HCI.

[8]  Carolina Cruz-Neira,et al.  Surround-Screen Projection-Based Virtual Reality: The Design and Implementation of the CAVE , 2023 .

[9]  Bruce H. Thomas,et al.  Supporting remote tangible robotic entertainment , 2008, ACE '08.

[10]  Elise van den Hoven,et al.  Tangible Computing in Everyday Life: Extending Current Frameworks for Tangible User Interfaces with Personal Objects , 2004, EUSAI.

[11]  Adam Kumpf,et al.  Trackmate : large-scale accessibility of Tangible User Interfaces , 2009 .

[12]  Scott R. Klemmer,et al.  Papier-Mache: toolkit support for tangible input , 2004, CHI.

[13]  Kasper Hornbæk,et al.  mixiTUI: a tangible sequencer for electronic live performances , 2009, Tangible and Embedded Interaction.

[14]  Thomas Hermann,et al.  Sonification for Exploratory Data Analysis , 2002 .

[15]  Kasper Hornbæk,et al.  Tangible bots: interaction with active tangibles in tabletop interfaces , 2011, CHI.

[16]  Jefferson Y. Han Low-cost multi-touch sensing through frustrated total internal reflection , 2005, UIST.

[17]  Mi Jeong Kim,et al.  The impact of tangible user interfaces on spatial cognition during collaborative design , 2008 .

[18]  Jiao Jian-li Literature Review in Learning with Tangible Technologies , 2008 .

[19]  Jacob Buur,et al.  Getting a grip on tangible interaction: a framework on physical space and social interaction , 2006, CHI.

[20]  Hirokazu Kato,et al.  Marker tracking and HMD calibration for a video-based augmented reality conferencing system , 1999, Proceedings 2nd IEEE and ACM International Workshop on Augmented Reality (IWAR'99).

[21]  Jennifer L. Gregg,et al.  The Networked Minds Measure of Social Presence : Pilot Test of the Factor Structure and Concurrent Validity , 2001 .

[22]  Saul Greenberg,et al.  Phidgets: easy development of physical interfaces through physical widgets , 2001, UIST '01.

[23]  Eckard Riedenklau TAOs - Tangible Active Objects for table-top interaction , 2009 .

[24]  H. Ritter,et al.  The cognitive nature of action - functional links between cognitive psychology, movement science, and robotics. , 2009, Progress in brain research.

[25]  Tovi Grossman,et al.  Medusa: a proximity-aware multi-touch tabletop , 2011, UIST.

[26]  Jean Rene Dawin,et al.  TRecS: Ein tangibles, rekonfigurierbares System zur explorativen Datenanalyse , 2008 .

[27]  Alan F. Blackwell,et al.  Peripheral tangible interaction by analytic design , 2009, Tangible and Embedded Interaction.

[28]  Yang Li,et al.  Protractor: a fast and accurate gesture recognizer , 2010, CHI.

[29]  Elise van den Hoven,et al.  Grasping gestures: Gesturing with physical artifacts , 2011, Artificial Intelligence for Engineering Design, Analysis and Manufacturing.

[30]  Eva Cerezo,et al.  ToyVision: a toolkit for prototyping tabletop tangible games , 2012, EICS '12.

[31]  Gregory Kramer,et al.  Auditory Display: Sonification, Audification, And Auditory Interfaces , 1994 .

[32]  J. B. Brooke,et al.  SUS: A 'Quick and Dirty' Usability Scale , 1996 .

[33]  Enrico Costanza,et al.  TUIO: A Protocol for Table-Top Tangible User Interfaces , 2005 .

[34]  Kei Nakatsuma,et al.  Touchable holography , 2009, SIGGRAPH '09.

[35]  Till Bovermann Tangible auditory interfaces: combining auditory displays and tangible interfaces , 2009 .

[36]  Paulo Guerra RemoteBunnies: multi-agent phenomena mapping between physical environments , 2013, TEI '13.

[37]  Li-Wei Chan,et al.  TUIC: enabling tangible interaction on capacitive multi-touch displays , 2011, CHI.

[38]  Michael S. Bernstein,et al.  Reflective physical prototyping through integrated design, test, and analysis , 2006, UIST.

[39]  I.,et al.  Fitts' Law as a Research and Design Tool in Human-Computer Interaction , 1992, Hum. Comput. Interact..

[40]  Ray A. Jarvis,et al.  On the Identification of the Convex Hull of a Finite Set of Points in the Plane , 1973, Inf. Process. Lett..

[41]  Paul P. Maglio,et al.  On Distinguishing Epistemic from Pragmatic Action , 1994, Cogn. Sci..

[42]  Gerald Penn,et al.  Living with a Tabletop: Analysis and Observations of Long Term Office Use of a Multi-Touch Table , 2007, Second Annual IEEE International Workshop on Horizontal Interactive Human-Computer Systems (TABLETOP'07).

[43]  W. Buxton,et al.  A study in two-handed input , 1986, CHI '86.

[44]  H. Yanco,et al.  Analysis of natural gestures for controlling robot teams on multi-touch tabletop surfaces , 2009, ITS '09.

[45]  Hiroshi Ishii,et al.  Bricks: laying the foundations for graspable user interfaces , 1995, CHI '95.

[46]  Hiroshi Ishii,et al.  Mechanical constraints as computational constraints in tabletop tangible interfaces , 2007, CHI.

[47]  Alois Ferscha,et al.  Physical shortcuts for media remote controls , 2008, INTETAIN '08.

[48]  Mikael Fernström,et al.  THE BEST OF TWO WORLDS: RETRIEVING AND BROWSING , 2000 .

[49]  Tara M Madhyastha,et al.  A Portable system for data sonification , 1992 .

[50]  Hiroshi Ishii,et al.  Sensetable: a wireless object tracking platform for tangible user interfaces , 2001, CHI.

[51]  Robert J. K. Jacob,et al.  The perceptual structure of multidimensional input device selection , 1992, CHI.

[52]  Jan O. Borchers,et al.  Madgets: actuating widgets on interactive tabletops , 2010, UIST.

[53]  Ken Hinckley,et al.  A survey of design issues in spatial input , 1994, UIST '94.

[54]  Abdelsalam Helal,et al.  Atlas: A Service-Oriented Sensor Platform: Hardware and Middleware to Enable Programmable Pervasive Spaces , 2006, Proceedings. 2006 31st IEEE Conference on Local Computer Networks.

[55]  T. Flash,et al.  The coordination of arm movements: an experimentally confirmed mathematical model , 1985, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[56]  Kenneth P. Fishkin,et al.  A taxonomy for and analysis of tangible interfaces , 2004, Personal and Ubiquitous Computing.

[57]  Nelson H. C. Yung,et al.  Corner detector based on global and local curvature properties , 2008 .

[58]  Hiroshi Ishii,et al.  Interaction techniques for musical performance with tabletop tangible interfaces , 2006, ACE '06.

[59]  Daisuke Sakamoto,et al.  CRISTAL: a collaborative home media and device controller based on a multi-touch display , 2009, ITS '09.

[60]  Ian Oakley,et al.  Haptics for tangible interaction: a vibro-tactile prototype , 2011, Tangible and Embedded Interaction.

[61]  Stavros N. Demetriadis,et al.  Tangible versus graphical user interfaces for robot programming: exploring cross-age children’s preferences , 2013, Personal and Ubiquitous Computing.

[62]  Helge Ritter,et al.  Supplementary Material for "Embodied Social Networking with Gesture-enabled Tangible Active Objects" , 2011 .

[63]  Brad A. Myers,et al.  Maximizing the guessability of symbolic input , 2005, CHI Extended Abstracts.

[64]  Dieter Schmalstieg,et al.  ARToolKitPlus for Pose Trackin on Mobile Devices , 2007 .

[65]  T. M. Gehring,et al.  Family-system-test (FAST): A three dimensional approach to investigate family relationships , 1986, Child psychiatry and human development.

[66]  Helge J. Ritter,et al.  Gesture Desk - An Integrated Multi-modal Gestural Workplace for Sonification , 2003, Gesture Workshop.

[67]  Sebastian Wrede,et al.  A middleware for collaborative research in experimental robotics , 2011, 2011 IEEE/SICE International Symposium on System Integration (SII).

[68]  Illah R. Nourbakhsh,et al.  A survey of socially interactive robots , 2003, Robotics Auton. Syst..

[69]  Orit Shaer,et al.  Reality-based interaction: a framework for post-WIMP interfaces , 2008, CHI.

[70]  Sowmya Somanath,et al.  Integrating a robot in a tabletop reservoir engineering application , 2013, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[71]  Helge Ritter,et al.  A Tangible Environment for Ambient Data Representation , 2006 .

[72]  Joseph A. Paradiso,et al.  Electromagnetic Tagging for Electronic Music Interfaces , 2003 .

[73]  Luís Magalhães,et al.  Different interaction paradigms for different user groups: an evaluation regarding content selection , 2014, Interacción '14.

[74]  Yvonne Rogers,et al.  A framework for designing sensor-based interactions to promote exploration and reflection in play , 2006, Int. J. Hum. Comput. Stud..

[75]  J. Underkoff Urp : A Luminous-Tangible Workbench for Urban Planning and Design , 1999, CHI 1999.

[76]  Teresa Marrin,et al.  Possibilities for the digital baton as a general-purpose gestural interface , 1997, CHI 1997.

[77]  Michael Rohs,et al.  A $3 gesture recognizer: simple gesture recognition for devices equipped with 3D acceleration sensors , 2010, IUI '10.

[78]  Niels Henze,et al.  Self-Actuated Displays for Vertical Surfaces , 2015, INTERACT.

[79]  Rex B. Kline,et al.  Becoming a Behavioral Science Researcher: A Guide to Producing Research That Matters , 2008 .

[80]  Karsten Nebe,et al.  dSensingNI: a framework for advanced tangible interaction using a depth camera , 2012, TEI.

[81]  Nicholas P. Holmes,et al.  The body schema and multisensory representation(s) of peripersonal space , 2004, Cognitive Processing.

[82]  Michael S. Horn,et al.  Comparing the use of tangible and graphical programming languages for informal science education , 2009, CHI.

[83]  Darren Leigh,et al.  The calder toolkit: wired and wireless components for rapidly prototyping interactive devices , 2004, DIS '04.

[84]  Jiajie Zhang,et al.  Representations in Distributed Cognitive Tasks , 1994, Cogn. Sci..

[85]  Meredith Ringel Morris,et al.  iStuff: a physical user interface toolkit for ubiquitous computing environments , 2003, CHI '03.

[86]  Ali Momeni,et al.  OpenSound Control: State of the Art 2003 , 2003, NIME.

[87]  Taylor Cl,et al.  The anatomy and mechanics of the human hand. , 1955, Artificial limbs.

[88]  Ehud Sharlin,et al.  On tangible user interfaces, humans and spatiality , 2004, Personal and Ubiquitous Computing.

[89]  Nick Collins,et al.  The SuperCollider Book , 2011 .

[90]  Yang Li,et al.  Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes , 2007, UIST.

[91]  H. Kuhn The Hungarian method for the assignment problem , 1955 .

[92]  M. Sheelagh T. Carpendale,et al.  Territoriality in collaborative tabletop workspaces , 2004, CSCW.

[93]  Stephan Wensveen,et al.  Interaction frogger: a design framework to couple action and function through feedback and feedforward , 2004, DIS '04.

[94]  G. Schlesinger Der mechanische Aufbau der künstlichen Glieder , 1919 .

[95]  James D. Hollan,et al.  SLAP widgets: bridging the gap between virtual and physical controls on tabletops , 2009, CHI.

[96]  Sriram Subramanian,et al.  Ultra-tangibles: creating movable tangible objects on interactive tables , 2012, CHI.

[97]  Alain Berthoz,et al.  Complex unconstrained three-dimensional hand movement and constant equi-affine speed. , 2009, Journal of neurophysiology.

[98]  Orit Shaer,et al.  The TAC paradigm: specifying tangible user interfaces , 2004, Personal and Ubiquitous Computing.

[99]  Bruce H. Thomas,et al.  Through-Walls Collaboration , 2009, IEEE Pervasive Computing.

[100]  Helge Ritter,et al.  TANGIBLE ACTIVE OBJECTS AND INTERACTIVE SONIFICATION AS A SCATTER PLOT ALTERNATIVE FOR THE VISUALLY IMPAIRED , 2010, ICAD 2010.

[101]  Alexander Toet,et al.  Subjective User Experience and Performance with Active Tangibles on a Tabletop Interface , 2015, HCI.

[102]  Masahiko Inami,et al.  Augmented coliseum: an augmented game environment with small vehicles , 2006, First IEEE International Workshop on Horizontal Interactive Human-Computer Systems (TABLETOP '06).

[103]  Helge J. Ritter,et al.  An integrated multi-modal actuated tangible user interface for distributed collaborative planning , 2012, Tangible and Embedded Interaction.

[104]  Helge Ritter,et al.  AudioDB: Get in Touch with Sounds , 2008 .

[105]  Gerd Kortuem,et al.  Affordance-Based Design of Physical Interfaces for Ubiquitous Environments , 2006, UCS.

[106]  Harald Reiterer,et al.  Adaptive personal territories for co-located tabletop interaction in a museum setting , 2011, ITS '11.

[107]  Christian Leichsenring,et al.  Turtledove: a tangible grain interface for image organization , 2012, CHI EA '12.

[108]  Elena Mugellini,et al.  Move, hold and touch : a framework for tangible gesture interactive systems , 2015 .

[109]  Yang Li,et al.  Gesture coder: a tool for programming multi-touch gestures by demonstration , 2012, CHI.

[110]  Masahiko Inami,et al.  Remote active tangible interactions , 2007, TEI.

[111]  Patrick Jermann,et al.  An empirical evaluation of touch and tangible interfaces for tabletop displays , 2010, TEI '10.

[112]  Masanori Sugimoto,et al.  RoboTable: a tabletop framework for tangible interaction with robots in a mixed reality , 2009, Advances in Computer Entertainment Technology.

[113]  Luca Benini,et al.  Continuous Gesture Recognition for Resource Constrained Smart Objects , 2010 .

[114]  Fernando Garcia-Sanjuan,et al.  TangiWheel: A Widget for Manipulating Collections on Tabletop Displays Supporting Hybrid Input Modality , 2012, Journal of Computer Science and Technology.

[115]  Lakshmi Kurup,et al.  Study of a Home Robot: JIBO , 2014 .

[116]  Helge Ritter,et al.  TANGIBLE COMPUTING FOR INTERACTIVE SONIFICATION OF MULTIVARIATE DATA , 2007 .

[117]  Alissa Nicole Antle,et al.  The CTI framework: informing the design of tangible systems for children , 2007, TEI.

[118]  Frank Maurer,et al.  A domain specific language to define gestures for multi-touch applications , 2010, DSM '10.

[119]  Helge J. Ritter,et al.  Approaching Manual Intelligence , 2010, KI - Künstliche Intelligenz.

[120]  Meredith Ringel Morris,et al.  Understanding users' preferences for surface gestures , 2010, Graphics Interface.

[121]  Tom Igoe,et al.  Physical computing: sensing and controlling the physical world with computers , 2004 .

[122]  Hiroshi Ishii,et al.  Tangible bits: towards seamless interfaces between people, bits and atoms , 1997, CHI.

[123]  Hiroshi Ishii,et al.  A tangible interface for organizing information using a grid , 2002, CHI.

[124]  Masahiko Inami,et al.  Active Tangible Interactions , 2010, Tabletops.

[125]  Meredith Ringel Morris,et al.  User-defined gestures for surface computing , 2009, CHI.

[126]  Mark Fiala,et al.  ARTag, a fiducial marker system using digital techniques , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[127]  Christian Heath,et al.  Embodied reference: A study of deixis in workplace interaction , 2000 .

[128]  D. A. Kenny,et al.  Enhanced co-orientation in the perception of friends: a social relations analysis. , 1994, Journal of personality and social psychology.

[129]  Stan Sclaroff,et al.  A Unified Framework for Gesture Recognition and Spatiotemporal Gesture Segmentation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[130]  R. Bencina,et al.  Improved Topological Fiducial Tracking in the reacTIVision System , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[131]  Cynthia Breazeal,et al.  Toward sociable robots , 2003, Robotics Auton. Syst..

[132]  Diego López-de-Ipiña,et al.  TRIP: A Low-Cost Vision-Based Location System for Ubiquitous Computing , 2002, Personal and Ubiquitous Computing.

[133]  Sebastian Wrede,et al.  An Integration Framework for Developing Interactive Robots , 2005, PPSDR@ICRA.

[134]  Thomas Hermann,et al.  TUImod: Modular Objects for Tangible User Interfaces , 2008 .

[135]  Ali Mazalek,et al.  Exploring the design space of gestural interaction with active tokens through user-defined gestures , 2014, CHI.

[136]  Patrick Reuter,et al.  GeoTUI: a tangible user interface for geoscience , 2008, TEI.

[137]  Yvonne Rogers,et al.  Collaboration and interference: awareness with mice or touch input , 2008, CSCW.

[138]  Massimo Banzi,et al.  Getting Started with Arduino , 2008 .

[139]  K. Kuutti Activity theory as a potential framework for human-computer interaction research , 1995 .

[140]  Lisa Anthony,et al.  A lightweight multistroke recognizer for user interface prototypes , 2010, Graphics Interface.

[141]  D. Lloyd,et al.  The space between us: A neurophilosophical framework for the investigation of human interpersonal space , 2009, Neuroscience & Biobehavioral Reviews.

[142]  Jakob Tholander,et al.  Beyond representations: towards an action-centric perspective on tangible interaction , 2008, Int. J. Arts Technol..

[143]  J. Fischer,et al.  The Prehensile Movements of the Human Hand , 2014 .

[144]  Bruce Blumberg,et al.  Sympathetic interfaces: using a plush toy to direct synthetic characters , 1999, CHI '99.

[145]  Hiroshi Ishii,et al.  mediaBlocks: physical containers, transports, and controls for online media , 1998, SIGGRAPH.

[146]  P. Fitts The information capacity of the human motor system in controlling the amplitude of movement. , 1954, Journal of experimental psychology.

[147]  Jay Seitz The bodily basis of thought , 2000 .

[148]  Satoru Miyano,et al.  Open source clustering software , 2004 .

[149]  Michael Rohs,et al.  Protractor3D: a closed-form solution to rotation-invariant 3D gestures , 2011, IUI '11.

[150]  A. Kendon Conducting Interaction: Patterns of Behavior in Focused Encounters , 1990 .

[151]  Hans-Werner Gellersen,et al.  A malleable control structure for softwired user interfaces , 2007, Tangible and Embedded Interaction.

[152]  Steve Benford,et al.  A Framework for Tangible User Interfaces , 2003 .

[153]  Aaron M. Dollar,et al.  Classifying human manipulation behavior , 2011, 2011 IEEE International Conference on Rehabilitation Robotics.

[154]  Ying Liu,et al.  Tactile gesture recognition for people with disabilities , 2005, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005..

[155]  John Moody,et al.  Fast Learning in Networks of Locally-Tuned Processing Units , 1989, Neural Computation.

[156]  Gabriel Bendersky,et al.  Pervasive Games , 2015, EAI Endorsed Trans. Serious Games.

[157]  Helge J. Ritter,et al.  Begreifbare sich bewegende Objekte in Tisch-basierten Interaktionsszenarien , 2012, i-com.

[158]  Sergi Jordà,et al.  The reacTable: exploring the synergy between live music performance and tabletop tangible interfaces , 2007, TEI.

[159]  Jiajie Zhang,et al.  The Nature of External Representations in Problem Solving , 1997, Cogn. Sci..

[160]  Alan F. Blackwell,et al.  Correlates of the cognitive dimensions for tangible user interface , 2006, J. Vis. Lang. Comput..

[161]  Philip Tuddenham,et al.  Graspables revisited: multi-touch vs. tangible input for tabletop displays in acquisition and manipulation tasks , 2010, CHI.

[162]  E. Erdfelder,et al.  Statistical power analyses using G*Power 3.1: Tests for correlation and regression analyses , 2009, Behavior research methods.

[163]  Hiroshi Ishii,et al.  Audiopad: A Tag-based Interface for Musical Performance , 2002, NIME.

[164]  Ross Bencina,et al.  reacTIVision: a computer-vision framework for table-based tangible interaction , 2007, TEI.

[165]  Heinz Wörn,et al.  Introducing Wanda - A new robot for research, education, and Arts , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[166]  Y. Guiard Asymmetric division of labor in human skilled bimanual action: the kinematic chain as a model. , 1987, Journal of motor behavior.

[167]  F. Previc The neuropsychology of 3-D space. , 1998, Psychological bulletin.

[168]  F. Previc Functional specialization in the lower and upper visual fields in humans: Its ecological origins and neurophysiological implications , 1990, Behavioral and Brain Sciences.

[169]  Paul Dourish,et al.  Where the action is , 2001 .

[170]  W. Buxton Human-Computer Interaction , 1988, Springer Berlin Heidelberg.

[171]  Abigail Sellen,et al.  At home with surface computing? , 2012, CHI.

[172]  Thomas Kienzl,et al.  MRI: a mixed reality interface for the masses , 2006, SIGGRAPH '06.

[173]  Austin Henderson,et al.  Making sense of sensing systems: five questions for designers and researchers , 2002, CHI.

[174]  Ipke Wachsmuth,et al.  From Body Space to Interaction Space - Modeling Spatial Cooperation for Virtual Humans , 2011 .

[175]  Tony P. Pridmore,et al.  Expected, sensed, and desired: A framework for designing sensing-based interaction , 2005, TCHI.

[176]  William Buxton,et al.  An empirical evaluation of graspable user interfaces: towards specialized, space-multiplexed input , 1997, CHI.

[177]  Orit Shaer,et al.  A specification paradigm for the design and implementation of tangible user interfaces , 2009, TCHI.

[178]  Sowmya Somanath,et al.  Spidey: a Robotic Tabletop Assistant , 2012 .

[179]  D. Maynes-Aminzade,et al.  The actuated workbench: computer-controlled actuation in tabletop tangible interfaces , 2003, ACM Trans. Graph..

[180]  Thomas Kienzl,et al.  tangible workbench "TW": with changeable markers , 2008, SIGGRAPH '08.

[181]  Ali Mazalek,et al.  Tangible Meets Gestural : Gesture Based Interaction with Active Tokens , 2014 .

[182]  Susan Goldin Hearing gesture : how our hands help us think , 2003 .

[183]  Anton Nijholt,et al.  Smart material interfaces: a new form of physical interaction , 2012, CHI EA '12.

[184]  Scott Brenner Brave Tangible interfaces for remote communication and collaboration , 1998 .

[185]  Scott R. Klemmer,et al.  How bodies matter: five themes for interaction design , 2006, DIS '06.

[186]  Abigail Sellen,et al.  Putting the physical into the digital: issues in designing hybrid interactive surfaces , 2009 .

[187]  Holger Regenbrecht,et al.  Social presence in two- and three-dimensional videoconferencing , 2005 .

[188]  Alexander Toet,et al.  Uni-, bi- and tri-modal warning signals: effects of temporal parameters and sensory modality on perceived urgency , 2015 .

[189]  E. Schegloff,et al.  A simplest systematics for the organization of turn-taking for conversation , 2015 .

[190]  Richard T. Vaughan,et al.  The Player/Stage Project: Tools for Multi-Robot and Distributed Sensor Systems , 2003 .

[191]  Hiroshi Ishii,et al.  Picture this!: film assembly using toy gestures , 2008, UbiComp.

[192]  T. Nakajima,et al.  CookieFlavors: Rapid Composition Framework for Tangible Media , 2007, The 2007 International Conference on Next Generation Mobile Applications, Services and Technologies (NGMAST 2007).

[193]  Hiroshi Ishii,et al.  Emerging frameworks for tangible user interfaces , 2000, IBM Syst. J..

[194]  Timo Arnall,et al.  Designing with RFID , 2009, Tangible and Embedded Interaction.

[195]  Hiroshi Ishii,et al.  SpaceTop: integrating 2D and spatial 3D interactions in a see-through desktop environment , 2013, CHI.

[196]  Meredith Ringel Morris,et al.  ShadowGuides: visualizations for in-situ learning of multi-touch and whole-hand gestures , 2009, ITS '09.

[197]  Radhika Nagpal,et al.  Kilobot: A low cost scalable robot system for collective behaviors , 2012, 2012 IEEE International Conference on Robotics and Automation.

[198]  Sergi Jordà,et al.  On stage: the reactable and other musical tangibles go real , 2008, Int. J. Arts Technol..

[199]  Ken Perlin,et al.  Physical objects as bidirectional user interface elements , 2004, IEEE Computer Graphics and Applications.

[200]  Hiroshi Ishii,et al.  Token+constraint systems for tangible interaction with digital information , 2005, TCHI.

[201]  Dan Overholt,et al.  Towards an Interface for Music Mixing based on Smart Tangibles and Multitouch , 2013, NIME.

[202]  Elise van den Hoven,et al.  Framing tangible interaction frameworks , 2009, Artificial Intelligence for Engineering Design, Analysis and Manufacturing.

[203]  Brygg Ullmer,et al.  Tangible interfaces for manipulating aggregates of digital information , 2002 .

[204]  Emily Sun,et al.  Sifteo cubes , 2012, CHI Extended Abstracts.

[205]  Tom Igoe Making Things Talk - Practical Methods for Connecting Physical Objects, Second Edition , 2011 .

[206]  Majken Kirkegaard Rasmussen,et al.  Shape-changing interfaces: a review of the design space and open research questions , 2012, CHI.

[207]  Hiroshi Ishii,et al.  The metaDESK: models and prototypes for tangible user interfaces , 1997, UIST '97.