Interaction with eyes-free and gestural interfaces

Aalto University, P.O. Box 11000, FI-00076 Aalto www.aalto.fi Author Raine Kajastila Name of the doctoral dissertation Interaction with eyes-free and gestural interfaces Publisher School of Science Unit Department of Media Technology Series Aalto University publication series DOCTORAL DISSERTATIONS 22/2013 Field of research Media technology Manuscript submitted 11 September 2012 Date of the defence 8 February 2013 Permission to publish granted (date) 17 December 2012 Language English Monograph Article dissertation (summary + original articles) Abstract Eyes-free interaction aims to control devices without the need to look at them. This is especially useful while driving, walking on a bustling street, or in other situations when looking at a display would be dangerous, inconvenient or restricted. Hand gestures and feedback with sound offer an eyes-free alternative to visual displays, and this thesis studies using them with devices and the surrounding environment.Eyes-free interaction aims to control devices without the need to look at them. This is especially useful while driving, walking on a bustling street, or in other situations when looking at a display would be dangerous, inconvenient or restricted. Hand gestures and feedback with sound offer an eyes-free alternative to visual displays, and this thesis studies using them with devices and the surrounding environment. In this thesis work, advanced circular auditory menus and three parallel control methods for using them were developed. Essentially, the thesis work concentrated on a circular interaction metaphor in auditory menus, in which the gesture was mapped directly to the position in the menu. The introduced control methods and auditory menu properties were tested with user experiments, and a mobile application integrating auditory and visual menus was built. The three gestural control methods to control circular auditory menus included accelerometer-based, touch screen-based, and camera-based interaction. All control methods were proven accurate and fast enough for efficient eyes-free use. Additionally, the same control methods were used in both visual and auditory domains, which facilitates switching to eyesfree use when needed and may also improve the accessibility of the interface for visually impaired users. Results of user experiments showed that the introduced visual and auditory menu design was easy and intuitive to learn without extensive training. Furthermore, a solution for eyes-free access to large menus was proposed, and user experiments indicated that dynamic menu item placement is efficient, accurate, and allowed the use of large menus. This thesis also investigated the use of auditory displays and gesture interfaces in performing arts. The perceived shape and size of a space can be changed by applying different reverberation times in different directions using multiple reverberation systems. Implementing a reverberation system and a test setup for subjective evaluation validated this. The implemented reverberation system has been utilized in live opera performances and to enhance lecture room acoustics. The use of gesture control is explored in an experimental opera production in which the performers controlled an audiovisual virtual stage live. The live interaction was useful when user controlled media was directly mapped onto gestures and when detailed nuances of movement were hard for a technician controlling the media to follow.

[1]  David G. Malham,et al.  3-D Sound Spatialization using Ambisonic Techniques , 1995 .

[2]  Pattie Maes,et al.  Intimate interfaces in action: assessing the usability and subtlety of emg-based motionless gestures , 2007, CHI.

[3]  Ville Pulkki,et al.  HRTF Measurements with a Continuously Moving Loudspeaker and Swept Sines , 2010 .

[4]  Jani Mäntyjärvi,et al.  Accelerometer-based gesture control for a design environment , 2006, Personal and Ubiquitous Computing.

[5]  Jacob O. Wobbrock,et al.  Slide rule: making mobile touch screens accessible to blind people using multi-touch interaction techniques , 2008, Assets '08.

[6]  Tapio Lokki,et al.  Augmented reality audio for mobile and wearable appliances , 2004 .

[7]  D. McAlpine,et al.  Neural mechanisms of binaural hearing , 2002 .

[8]  Alex Pentland,et al.  Augmented Performance in Dance and Theater , 1999 .

[9]  Marcelo M. Wanderley,et al.  Trends in Gestural Control of Music , 2000 .

[10]  Marinus M. Boone,et al.  Spatial sound-field reproduction by wave-field synthesis , 1995 .

[11]  Meredith Ringel Morris,et al.  User-defined gestures for surface computing , 2009, CHI.

[12]  J. B. Brooke,et al.  SUS: A 'Quick and Dirty' Usability Scale , 1996 .

[13]  Peter Fröhlich,et al.  Markerless visual fingertip detection for natural mobile device interaction , 2011, Mobile HCI.

[14]  Angelo Farina,et al.  Simultaneous Measurement of Impulse Response and Distortion with a Swept-Sine Technique , 2000 .

[15]  J. Blauert Spatial Hearing: The Psychophysics of Human Sound Localization , 1983 .

[16]  Hans Wallach,et al.  The Precedence Effect in Sound Localization (Tutorial Reprint) , 1973 .

[17]  Ian Oakley,et al.  Designing Eyes-Free Interaction , 2007, HAID.

[18]  Simone Spagnol,et al.  Fitting pinna-related transfer functions to anthropometry for binaural sound rendering , 2010, 2010 IEEE International Workshop on Multimedia Signal Processing.

[19]  Maggie Orth Interface to architecture: integrating technology into the environment in the Brain Opera , 1997, DIS '97.

[20]  Shumin Zhai,et al.  The benefits of augmenting telephone voice menu navigation with visual browsing and search , 2006, CHI.

[21]  I. Scott MacKenzie,et al.  Reducing visual demand for gestural text input on touchscreen devices , 2012, CHI EA '12.

[22]  Lorna M. Brown,et al.  Tactile feedback for mobile interactions , 2007, CHI.

[23]  Mario Romero,et al.  BrailleTouch: Mobile Texting for the Visually Impaired , 2011, HCI.

[24]  Stephen A. Brewster,et al.  Audio or tactile feedback: which modality when? , 2009, CHI.

[25]  Erik Sikström,et al.  Designing Auditory Display Menu Interfaces - Cues for Users Current Location in Extensive Menus , 2009 .

[26]  Mark Billinghurst,et al.  A user study of auditory versus visual interfaces for use while driving , 2008, Int. J. Hum. Comput. Stud..

[27]  Vibha Sazawal,et al.  The Unigesture Approach , 2002, Mobile HCI.

[28]  Maurizio Gentilucci,et al.  On gesture and speech , 2015 .

[29]  Florian Mueller,et al.  Transparent hearing , 2002, CHI Extended Abstracts.

[30]  Richard A. Bolt,et al.  “Put-that-there”: Voice and gesture at the graphics interface , 1980, SIGGRAPH '80.

[31]  Ian Oakley,et al.  Auditory display design for exploration in mobile audio-augmented reality , 2012, Personal and Ubiquitous Computing.

[32]  Stephen A. Brewster,et al.  Wrist rotation for interaction in mobile contexts , 2008, Mobile HCI.

[33]  Rainer Groh,et al.  Towards a formalization of multi-touch gestures , 2010, ITS '10.

[34]  Roy Want,et al.  Squeeze me, hold me, tilt me! An exploration of manipulative user interfaces , 1998, CHI.

[35]  Durand R. Begault,et al.  Perceptual Effects of Synthetic Reverberation on Three-Dimensional Audio Systems , 1992 .

[36]  Myounghoon Jeon,et al.  Ergonomics Society of the Human Factors and Human Factors: The Journal , 2012 .

[37]  Pierre Dragicevic,et al.  Earpod: eyes-free menu selection using touch input and reactive audio feedback , 2007, CHI.

[38]  Matti Karjalainen,et al.  Virtual Air Guitar , 2006 .

[39]  Ville Pulkki,et al.  Spatial sound generation and perception by amplitude panning techniques , 2001 .

[40]  D R Begault,et al.  Virtual acoustics, aeronautics, and communications. , 1998, Journal of the Audio Engineering Society. Audio Engineering Society.

[41]  Kent Lyons,et al.  The Gesture Watch: A Wireless Contact-free Gesture based Wrist Interface , 2007, 2007 11th IEEE International Symposium on Wearable Computers.

[42]  Michael Cohen,et al.  Multidimensional Audio Window Management , 1991, Int. J. Man Mach. Stud..

[43]  Timo Peltonen,et al.  A Rehearsal Hall with Virtual Acoustics for Symphony Orchestras , 2009 .

[44]  Myounghoon Jeon,et al.  Auditory menus are not just spoken visual menus: a case study of "unavailable" menu items , 2010, CHI Extended Abstracts.

[45]  Stephen Brewster,et al.  Extending the Auditory Display Space in Handheld Computing Devices , 1999 .

[46]  John Bowers,et al.  Assembling the senses: towards the design of cooperative interfaces for visually impaired users , 2004, CSCW.

[47]  William W. Gaver Auditory Icons: Using Sound in Computer Interfaces , 1986, Hum. Comput. Interact..

[48]  Ravin Balakrishnan,et al.  User learning and performance with bezel menus , 2012, CHI.

[49]  I. Scott MacKenzie,et al.  Eyes-free text entry with error correction on touchscreen mobile devices , 2010, NordiCHI.

[50]  Myounghoon Jeon,et al.  “Spindex”: Accelerated Initial Speech Sounds Improve Navigation Performance in Auditory Menus , 2009 .

[51]  Niels Henze,et al.  User-centred process for the definition of free-hand gestures applied to controlling music playback , 2012, Multimedia Systems.

[52]  Gregory D. Abowd,et al.  No-Look Notes: Accessible Eyes-Free Multi-touch Text Entry , 2010, Pervasive.

[53]  Tovi Grossman,et al.  Implanted user interfaces , 2012, CHI.

[54]  Chris Schmandt,et al.  Dynamic Soundscape: mapping time to space for audio browsing , 1997, CHI.

[55]  J. P. Lewis,et al.  SmartCanvas: a gesture-driven intelligent drawing desk system , 2005, IUI.

[56]  Khai N. Truong,et al.  Virtual shelves: interactions with orientation aware devices , 2009, UIST '09.

[57]  Daniel J. Wigdor,et al.  TiltText: using tilt for text input to mobile phones , 2003, UIST '03.

[58]  Marilyn Tremaine,et al.  Bullseye! when Fitts' law doesn't fit , 1998, CHI '98.

[59]  Constantine Stephanidis,et al.  A generic direct-manipulation 3D-auditory environment for hierarchical navigation in non-visual interaction , 1996, Assets '96.

[60]  Keith J. Burnham,et al.  A Research Study of Hand Gesture Recognition Technologies and Applications for Human Vehicle Interaction , 2007 .

[61]  Christopher Frauenberger,et al.  Mode independent interaction pattern design , 2005, Ninth International Conference on Information Visualisation (IV'05).

[62]  Gilles Bailly,et al.  Gesture-aware remote controls: guidelines and interaction technique , 2011, ICMI '11.

[63]  Christopher Frauenberger,et al.  Interaction patterns for auditory user interfaces , 2005 .

[64]  Bruce N. Walker,et al.  SPEARCONS: SPEECH-BASED EARCONS IMPROVE NAVIGATION PERFORMANCE IN AUDITORY MENUS , 2006 .

[65]  Vibha Sazawal,et al.  TiltType: accelerometer-supported text entry for very small devices , 2002, UIST '02.

[66]  Jeremy R. Cooperstock,et al.  A Paradigm for Physical Interaction with Sound in 3-D Audio Space , 2006, ICMC.

[67]  Amy Neustein Advances in Speech Recognition: Mobile Environments, Call Centers and Clinics , 2010 .

[68]  Stephen A. Brewster,et al.  Multimodal 'eyes-free' interaction techniques for wearable devices , 2003, CHI '03.

[69]  BillinghurstMark,et al.  A user study of auditory versus visual interfaces for use while driving , 2008 .

[70]  Stephen A. Brewster,et al.  Eyes-free multitasking: the effect of cognitive load on mobile spatial audio interfaces , 2011, CHI.

[71]  A. Pentland,et al.  Computer Vision for Human–Machine Interaction: A Framework for Gesture Generation and Interpretation , 1998 .

[72]  Gaetan Lorho,et al.  FEASIBILITY OF MULTIPLE NON-SPEECH SOUNDS PRESENTATION USING HEADPHONES , 2001 .

[73]  Morten Fjeld,et al.  Exploring user motivations for eyes-free interaction on mobile devices , 2012, CHI.

[74]  Leonello Tarabella,et al.  Handel, a Free-Hands Gesture Recognition System , 2004, CMMR.

[75]  Chris Schmandt,et al.  Nomadic radio: speech and audio interaction for contextual messaging in nomadic environments , 2000, TCHI.

[76]  Jonathan Grudin,et al.  Computer Supported Cooperative Work and Groupware , 1997, Adv. Comput..

[77]  Peter Robinson,et al.  The use of gestures in multimodal input , 1998, Assets '98.

[78]  Mark Levy,et al.  3D INTERACTIVE ENVIRONMENT FOR MUSIC COLLECTION NAVIGATION , 2008 .

[79]  Sriram Subramanian,et al.  Tilt techniques: investigating the dexterity of wrist-based input , 2009, CHI.

[80]  Marilyn Tremaine,et al.  A Bullseye Menu With Sound Feedback , 1997, HCI.

[81]  Jan O. Borchers,et al.  Tactile motion instructions for physical activities , 2009, CHI.

[82]  Myounghoon Jeon,et al.  Enhanced auditory menu cues improve dual task performance and are preferred with in-vehicle technologies , 2009, AutomotiveUI.

[83]  S Puckette Miller,et al.  Pure Data : another integrated computer music environment , 1996 .

[84]  Claudio S. Pinhanez,et al.  It/I: an experiment towards interactive theatrical performances , 1998, CHI Conference Summary.

[85]  Ikuko Eguchi Yairi,et al.  The evaluation of visually impaired people's ability of defining the object location on touch-screen , 2010, ASSETS '10.

[86]  Pai H. Chou,et al.  A wearable wireless sensor platform for interactive dance performances , 2006, Fourth Annual IEEE International Conference on Pervasive Computing and Communications (PERCOM'06).

[87]  Constantine Stephanidis,et al.  A 3D-auditory environment for hierarchical navigation in non-visual interaction , 1996 .

[88]  M Akamatsu,et al.  Please Scroll down for Article Ergonomics a Comparison of Tactile, Auditory, and Visual Feedback in a Pointing Task Using a Mouse-type Device , 2022 .

[89]  T. Lokki,et al.  An electro-acoustic enhancement system for rehearsal room , 2000 .

[90]  Gregory H. Wakefield,et al.  Introduction to Head-Related Transfer Functions (HRTFs): Representations of HRTFs in Time, Frequency, and Space , 2001 .

[91]  Durand R. Begault,et al.  3-D Sound for Virtual Reality and Multimedia Cambridge , 1994 .

[92]  Richard E. Ladner,et al.  Usable gestures for blind people: understanding preference and performance , 2011, CHI.

[93]  John G. Neuhoff,et al.  Sonification Report: Status of the Field and Research Agenda Prepared for the National Science Foundation by members of the International Community for Auditory Display , 1999 .

[94]  Meera Blattner,et al.  Earcons and Icons: Their Structure and Common Design Principles , 1989, Hum. Comput. Interact..

[95]  Joaquim A. Jorge,et al.  From Tapping to Touching: Making Touch Screens Accessible to Blind Users , 2008, IEEE MultiMedia.

[96]  Davide Rocchesso,et al.  Audio effects to enhance spatial information displays , 2002, Proceedings. First International Symposium on 3D Data Processing Visualization and Transmission.

[97]  M. Weiser,et al.  An empirical comparison of pie vs. linear menus , 1988, CHI '88.

[98]  Patrick Baudisch,et al.  Imaginary interfaces: spatial interaction with empty hands and without visual feedback , 2010, UIST.

[99]  Stephen Brewster,et al.  A Detailed Investigation into the Effectiveness of Earcons , 1997 .

[100]  Ramani Duraiswami,et al.  Extracting the frequencies of the pinna spectral notches in measured head related impulse responses. , 2004, The Journal of the Acoustical Society of America.

[101]  Gregory D. Abowd,et al.  BrailleTouch: designing a mobile eyes-free soft keyboard , 2011, Mobile HCI.

[102]  Helge J. Ritter,et al.  Gesture Desk - An Integrated Multi-modal Gestural Workplace for Sonification , 2003, Gesture Workshop.

[103]  Philip T. Kortum,et al.  Determining what individual SUS scores mean: adding an adjective rating scale , 2009 .

[104]  Thomas B. Moeslund,et al.  A procedure for developing intuitive and ergonomic gesture interfaces for man-machine interaction , 2003 .

[105]  Ian Oakley,et al.  A motion-based marking menu system , 2007, CHI Extended Abstracts.

[106]  Jun Rekimoto,et al.  Tilting operations for small screen interfaces , 1996, UIST '96.

[107]  Richard E. Ladner,et al.  V-braille: haptic braille perception using a touch-screen and vibration on mobile phones , 2010, ASSETS '10.

[108]  Rahul Sawhney,et al.  Sonic Grid: an auditory interface for the visually impaired to navigate GUI-based environments , 2008, IUI '08.

[109]  Helmut Haas,et al.  The Influence of a Single Echo on the Audibility of Speech , 1972 .

[110]  G. W. Furnas,et al.  Generalized fisheye views , 1986, CHI '86.

[111]  Stefanie Kuhn Extended presence: The instrumental(ised) body in André Werner's Marlowe: The Jew of Malta , 2007 .

[112]  Michael J. Gerzon Periphony: With-Height Sound Reproduction , 1973 .

[113]  Roderick Murray-Smith,et al.  Shoogle: excitatory multimodal interaction on mobile devices , 2007, CHI.

[114]  Stephen A. Brewster,et al.  Gestural and audio metaphors as a means of control for mobile devices , 2002, CHI.

[115]  R. Duda,et al.  Range dependence of the response of a spherical head model , 1998 .

[116]  Yael Edan,et al.  Vision-based hand-gesture applications , 2011, Commun. ACM.

[117]  Sean White,et al.  Nenya: subtle and eyes-free mobile input with a magnetically-tracked finger ring , 2011, CHI.

[118]  Katrin Wolf,et al.  Foogue: eyes-free interaction for smartphones , 2010, Mobile HCI.

[119]  mc schraefel,et al.  A Taxonomy of Gestures in Human Computer Interactions , 2005 .

[120]  David McGookin,et al.  Diary in the Sky: A Spatial Audio Display for a Mobile Calendar , 2001, BCS HCI/IHM.

[121]  Joseph A. Paradiso,et al.  The Brain Opera Technology: New Instruments and Gestural Sensors for Musical Interaction and Performance , 1999 .

[122]  Mark Apperley,et al.  Data base navigation: an office environment for the professional , 1982 .

[123]  Hongan Wang,et al.  Tilt menu: using the 3D orientation information of pen devices to extend the selection capability of pen-based user interfaces , 2008, CHI.

[124]  Stephen A. Brewster,et al.  Effects of feedback, mobility and index of difficulty on deictic spatial audio target acquisition in the horizontal plane , 2006, CHI.

[125]  Mark Billinghurst,et al.  Experiments in mobile spatial audio-conferencing: key-based and gesture-based interaction , 2008, Mobile HCI.

[126]  Shumin Zhai,et al.  SHRIMP: solving collision and out of vocabulary problems in mobile predictive input with motion gesture , 2010, CHI.