earPod: Efficient Hierarchical Eyes-free Menu Selection

The research in this dissertation developed and evaluated a new method for menuing interaction that is intended to be better suited than current methods with respect to mobile eyes-free scenarios. The earPod prototype was developed and then evaluated in a series of four experiments. In the first two experiments, earPod was first compared against an iPod-like (visual) interface and then against a fuller set of competitive techniques that included dual vs. single modality presentations, audio vs. visual modalities, and radial vs. linear mappings. The third experiment consisted of a longitudinal study designed to understand the learning patterns that occurred with these techniques. The fourth experiment examined performance in a conventional (single task) desktop setting and in a driving simulator (i.e., a dual task situation where participants carried out the driving task while interacting with the mobile device). The results of these experiments, comparing earPod with an iPod-like visual linear menu technique on fixed-sized static menus, indicated that earPod is comparable both in terms of speed and accuracy. Thus it seems likely that earPod should be an effective and efficient eyes-free menu selection technique. The comprehensive 3x2 study implemented in Experiment 2 showed that the benefit of earPod was largely due to the radial menu style design. While performance using it was comparable in both speed and accuracy with the visual linear menus, its performance was slower than for a visual radial style menu. In the multi-task simulated driving condition in Experiment 4, where concurrent tasks competed for visual attention, the eyes-free earPod interface was found to be beneficial in improving performance with respect to the safety related driving parameters of following distance and lateral movement in the lane. Thus auditory feedback appears to mitigate some of the risk associated with menu selection while driving. Overall, the results indicated that not only should earPod menuing be able to provide safer interaction in dual task settings, but also that, with sufficient training, audio only menu selection using innovative techniques such as those employed by earPod can be competitive with visual menuing systems even in desktop settings.

[1]  Ivan Poupyrev,et al.  Gummi: a bendable computer , 2004, CHI '04.

[2]  Brad A. Myers,et al.  EdgeWrite: a stylus-based text entry method designed for high accuracy and stability of motion , 2003, UIST '03.

[3]  Chris Schmandt,et al.  Nomadic radio: scaleable and contextual notification for wearable audio messaging , 1999, CHI '99.

[4]  Pierre Dragicevic,et al.  Earpod: eyes-free menu selection using touch input and reactive audio feedback , 2007, CHI.

[5]  Terry Winograd,et al.  FlowMenu: combining command, text, and data entry , 2000, UIST '00.

[6]  Ben Shneiderman,et al.  Split menus: effectively using selection frequency to organize menus , 1994, TCHI.

[7]  T. Landauer,et al.  Handbook of Human-Computer Interaction , 1997 .

[8]  Beryl Plimmer,et al.  Multimodal collaborative handwriting training for visually-impaired people , 2008, CHI.

[9]  William Buxton,et al.  The limits of expert performance using hierarchic marking menus , 1993, INTERCHI.

[10]  Dario D. Salvucci Predicting the effects of in-car interface use on driver performance: an integrated model approach , 2001, Int. J. Hum. Comput. Stud..

[11]  Desney S. Tan,et al.  InkSeine: In Situ search for active note taking , 2007, CHI.

[12]  Shumin Zhai,et al.  Hard lessons: effort-inducing interfaces benefit spatial learning , 2007, CHI.

[13]  David P. Morgan,et al.  How to build a speech recognition application : a style guide for telephony dialogues , 2001 .

[14]  Stephen A. Brewster,et al.  Gestural and audio metaphors as a means of control for mobile devices , 2002, CHI.

[15]  William Buxton,et al.  Tracking menus , 2003, UIST '03.

[16]  William Buxton,et al.  User learning and performance with marking menus , 1994, CHI 1994.

[17]  Kristen L. Macuga,et al.  Predicting the effects of cellular-phone dialing on driver performance , 2002, Cognitive Systems Research.

[18]  Susan J. Lederman,et al.  Design and Performance of a Tactile Shape Display Using RC Servomotors (Short Paper) , 2004 .

[19]  Jun Rekimoto,et al.  PreSense: interaction techniques for finger sensing input devices , 2003, UIST '03.

[20]  Stephen Brewster,et al.  Nonspeech auditory output , 2002 .

[21]  Elizabeth D. Mynatt,et al.  Nonvisual presentation of graphical user interfaces: contrasting two approaches , 1994, CHI Conference Companion.

[22]  Martin Hitz,et al.  Improving menu interaction: a comparison of standard, force enhanced and jumping menus , 2006, CHI.

[23]  Chris Schmandt,et al.  Nomadic radio: speech and audio interaction for contextual messaging in nomadic environments , 2000, TCHI.

[24]  W. E. Hick Quarterly Journal of Experimental Psychology , 1948, Nature.

[25]  James L. McClelland,et al.  The TRACE model of speech perception , 1986, Cognitive Psychology.

[26]  W. Marslen-Wilson Functional parallelism in spoken word-recognition , 1987, Cognition.

[27]  Gina-Anne Levow,et al.  Designing SpeechActs: issues in speech user interfaces , 1995, CHI '95.

[28]  Stephen A. Brewster,et al.  An evaluation of earcons for use in auditory human-computer interfaces , 1993, INTERCHI.

[29]  H Alm,et al.  Changes in driver behaviour as a function of handsfree mobile phones--a simulator study. , 1994, Accident; analysis and prevention.

[30]  John M. Flach,et al.  Control Theory for Humans: Quantitative Approaches To Modeling Performance , 2002 .

[31]  Gilles Bailly,et al.  Flower menus: a new type of marking menu with large menu breadth, within groups and efficient expert mode memorization , 2008, AVI '08.

[32]  Abigail Sellen,et al.  An Empirical Evaluation of Some Articulatory and Cognitive Aspects of Marking Menus , 1993, Hum. Comput. Interact..

[33]  Janan Al-Awar Smither,et al.  A Comparison of Broad Versus Deep Auditory Menu Structures , 2008, Hum. Factors.

[34]  Patrick Baudisch,et al.  Blindsight: eyes-free access to mobile phones , 2008, CHI.

[35]  P. Fitts The information capacity of the human motor system in controlling the amplitude of movement. , 1954, Journal of experimental psychology.

[36]  Emmanuel Barillot,et al.  Control menus: excecution and control in a single interactor , 2000, CHI Extended Abstracts.

[37]  K. Hinckley Input technologies and techniques , 2002 .

[38]  James D. Foley,et al.  The human factors of computer graphics interaction techniques , 1984, IEEE Computer Graphics and Applications.

[39]  R. Klatzky Human Memory: Structures And Processes , 1975 .

[40]  D. Strayer,et al.  Cell phone-induced failures of visual attention during simulated driving. , 2003, Journal of experimental psychology. Applied.

[41]  Barbara Freeman,et al.  Curing the menu blues in touch-tone voice interfaces , 2001, CHI Extended Abstracts.

[42]  Roy Want,et al.  Squeeze me, hold me, tilt me! An exploration of manipulative user interfaces , 1998, CHI.

[43]  Shumin Zhai,et al.  Quantifying coordination in multiple DOF movement and its application to evaluating 6 DOF input devices , 1998, CHI.

[44]  Virpi Roto,et al.  Interaction in 4-second bursts: the fragmented nature of attentional resources in mobile HCI , 2005, CHI.

[45]  Chris Schmandt,et al.  NewsComm: a hand-held interface for interactive access to structured audio , 1996, CHI.

[46]  I. Scott MacKenzie,et al.  Touchpad-based remote control devices , 1998, CHI Conference Summary.

[47]  Mike Sinclair,et al.  Touch-sensing input devices , 1999, CHI '99.

[48]  Marilyn Tremaine,et al.  Bullseye! when Fitts' law doesn't fit , 1998, CHI '98.

[49]  Hannu Korhonen Audio feedback on a mobile phone for number dialing , 2005, Mobile HCI.

[50]  Vincent Hayward,et al.  A role for haptics in mobile interaction: initial design using a handheld tactile display prototype , 2006, CHI.

[51]  Shumin Zhai,et al.  The benefits of augmenting telephone voice menu navigation with visual browsing and search , 2006, CHI.

[52]  William Buxton,et al.  Issues in combining marking and direct manipulation techniques , 1991, UIST '91.

[53]  Jakob Nielsen,et al.  Finding usability problems through heuristic evaluation , 1992, CHI.

[54]  T. V. Raman,et al.  Emacspeak—a speech interface , 1996, CHI.

[55]  William W. Gaver,et al.  AUDITORY ICONS IN LARGE-SCALE COLLABORATIVE ENVIRONMENTS , 1990, SGCH.

[56]  Ben Shneiderman,et al.  Designing the User Interface: Strategies for Effective Human-Computer Interaction , 1998 .

[57]  Barry Arons,et al.  SpeechSkimmer: a system for interactively skimming recorded speech , 1997, TCHI.

[58]  I. Scott MacKenzie,et al.  Graffiti vs. unistrokes: an empirical comparison , 2008, CHI.

[59]  Teresa L. Roberts,et al.  The effects of device technology on the usability of advanced telephone functions , 1989, CHI '89.

[60]  L. R. Peterson,et al.  Short-term retention of individual verbal items. , 1959, Journal of experimental psychology.

[61]  Stephen A. Brewster,et al.  Maximising screen-space on mobile computing devices , 1999, CHI Extended Abstracts.

[62]  Andrea R. Kennel Audiograf: a diagram-reader for the blind , 1996, Assets '96.

[63]  Ian Oakley,et al.  A motion-based marking menu system , 2007, CHI Extended Abstracts.

[64]  Marilyn A. Walker,et al.  What can I say?: evaluating a spoken language interface to Email , 1998, CHI.

[65]  Jennifer Balogh,et al.  Voice User Interface Design , 2004 .

[66]  G. A. Miller THE PSYCHOLOGICAL REVIEW THE MAGICAL NUMBER SEVEN, PLUS OR MINUS TWO: SOME LIMITS ON OUR CAPACITY FOR PROCESSING INFORMATION 1 , 1956 .

[67]  I. Scott MacKenzie,et al.  The tactile touchpad , 1997, CHI Extended Abstracts.

[68]  Ian H. Witten,et al.  On Frequency-Based Menu-Splitting Algorithms , 1984, Int. J. Man Mach. Stud..

[69]  Gilles Bailly,et al.  Wave Menus: Improving the Novice Mode of Hierarchical Marking Menus , 2007, INTERACT.

[70]  A J McKnight,et al.  The effect of cellular phone use upon driver attention. , 1993, Accident; analysis and prevention.

[71]  Allen and Rosenbloom Paul S. Newell,et al.  Mechanisms of Skill Acquisition and the Law of Practice , 1993 .

[72]  Duncan P. Brumby,et al.  iPod distraction: effects of portable music-player use on driver performance , 2007, CHI.

[73]  T. K. Landauer,et al.  Selection from alphabetic and numeric menu trees using a touch screen: breadth, depth, and width , 1985, CHI '85.

[74]  Mary Czerwinski,et al.  Web page design: implications of memory, structure and scent for information retrieval , 1998, CHI.

[75]  Shumin Zhai,et al.  SHARK2: a large vocabulary shorthand writing system for pen-based computers , 2004, UIST '04.

[76]  Ravin Balakrishnan,et al.  VisionWand: interaction techniques for large displays using a passive wand tracked in 3D , 2004, SIGGRAPH 2004.

[77]  Andy Cockburn,et al.  Faster cascading menu selections with enlarged activation areas , 2006, Graphics Interface.

[78]  B. Bederson,et al.  Understanding Single-Handed Mobile Device Interaction , 2006 .

[79]  James A. Landay,et al.  SATIN: a toolkit for informal ink-based applications , 2000, UIST '00.

[80]  Nicole Yankelovich,et al.  Designing the user interface for speech recognition applications , 1996, CHI 1996.

[81]  Mark H. Chignell,et al.  Searching in audio: the utility of transcripts, dichotic presentation, and time-compression , 2006, CHI.

[82]  J. Gibson Observations on active touch. , 1962, Psychological review.

[83]  Barry Arons,et al.  VoiceNotes: a speech interface for a hand-held voice notetaker , 1993, INTERCHI.

[84]  Stephen A. Brewster,et al.  Using nonspeech sounds to provide navigation cues , 1998, TCHI.

[85]  Meera Blattner,et al.  Earcons and Icons: Their Structure and Common Design Principles , 1989, Hum. Comput. Interact..

[86]  Dario D. Salvucci A Multitasking General Executive for Compound Continuous Tasks , 2005, Cogn. Sci..

[87]  Pierre Dragicevic,et al.  Strategies for accelerating on-line learning of hotkeys , 2007, CHI.

[88]  Philip R. Cohen,et al.  Synergistic use of direct manipulation and natural language , 1989, CHI '89.

[89]  Chunyuan Liao,et al.  Pen-top feedback for paper-based interfaces , 2006, UIST.

[90]  William W. Gaver The SonicFinder: An Interface That Uses Auditory Icons , 1989, Hum. Comput. Interact..

[91]  Maneesh Agrawala,et al.  Zone and polygon menus: using relative position to increase the breadth of multi-stroke marking menus , 2006, CHI.

[92]  Daniel J. Wigdor,et al.  TiltText: using tilt for text input to mobile phones , 2003, UIST '03.

[93]  Amy L. Schwartz,et al.  Increasing the Usability of Interactive Voice Response Systems: Research and Guidelines for Phone-Based Interfaces , 1995, Hum. Factors.

[94]  James A. Landay,et al.  Sketching images eyes-free: a grid-based dynamic drawing tool for the blind , 2002, Assets '02.

[95]  Grigore C. Burdea,et al.  Force and Touch Feedback for Virtual Reality , 1996 .

[96]  Joseph G. Pigeon,et al.  Statistics for Experimenters: Design, Innovation and Discovery , 2006, Technometrics.

[97]  Poulton Ec,et al.  Unwanted asymmetrical transfer effects with balanced experimental designs. , 1966 .

[98]  Dwight P. Miller The Depth/Breadth Tradeoff in Hierarchical Computer Menus , 1981 .

[99]  Dinesh K. Pai,et al.  JASS: A JAVA AUDIO SYNTHESIS SYSTEM FOR PROGRAMMERS , 2001 .

[100]  Gordon Kurtenbach,et al.  The design and evaluation of marking menus , 1993 .

[101]  William Buxton,et al.  Issues and techniques in touch-sensitive tablet input , 1985, SIGGRAPH '85.

[102]  Ravin Balakrishnan,et al.  Simple vs. compound mark hierarchical marking menus , 2004, UIST '04.

[103]  Alistair D. N. Edwards,et al.  Soundtrack: An Auditory Interface for Blind Users (Abstract Only) , 1989, SGCH.

[104]  Daniel Vogel,et al.  HybridPointing: fluid switching between absolute and relative pointing with a direct input device , 2006, UIST.

[105]  Ben Shneiderman,et al.  The Psychology of Menu Selection: Designing Cognitive Control at the Human/Computer Interface , 1991 .

[106]  David Goldberg,et al.  Touch-typing with a stylus , 1993, INTERCHI.

[107]  M. Weiser,et al.  An empirical comparison of pie vs. linear menus , 1988, CHI '88.

[108]  M. Atkins,et al.  Evaluating Interactive Technologies for Learning. , 1993 .

[109]  David L. Strayer,et al.  Driven to Distraction: Dual-Task Studies of Simulated Driving and Conversing on a Cellular Telephone , 2001, Psychological science.

[110]  Patrick Baudisch,et al.  Design and analysis of delimiters for selection-action pen gesture phrases in scriboli , 2005, CHI.

[111]  Shumin Zhai,et al.  Human Performance in Six Degree of Freedom Input Control , 2002 .

[112]  Christopher D. Wickens,et al.  Multiple resources and performance prediction , 2002 .

[113]  Stephen A. Brewster,et al.  Multimodal 'eyes-free' interaction techniques for wearable devices , 2003, CHI '03.

[114]  David R. Morse,et al.  Using while moving: HCI issues in fieldwork environments , 2000, TCHI.

[115]  R. Hyman Stimulus information as a determinant of reaction time. , 1953, Journal of experimental psychology.

[116]  William W. Gaver,et al.  Chapter 42 – Auditory Interfaces , 1997 .

[117]  Masanori Sugimoto,et al.  HybridTouch: an intuitive manipulation technique for PDAs using their front and rear surfaces , 2006, Mobile HCI.

[118]  Charlotte Magnusson,et al.  Iterative design of an audio-haptic drawing application , 2007, CHI Extended Abstracts.

[119]  Paul Resnick,et al.  Skip and scan: cleaning up telephone interface , 1992, CHI '92.

[120]  Mark W. Newman,et al.  DENIM: finding a tighter fit between tools and practice for Web site design , 2000, CHI.

[121]  George Engelbeck,et al.  Designing Voice Menu Applications for Telephones , 1997 .