Shared Input Multimodal Mobile Interfaces: Interaction Modality Effects on Menu Selection in Single-Task and Dual-Task Environments

Audio and visual modalities are two common output channels in the user interfaces embedded in today’s mobile devices. However, these user interfaces typically center on the visual modality as the primary output channel, with audio output serving a secondary role. This paper argues for an increased need for shared input multimodal user interfaces for mobile devices. A shared input multimodal interface can be operated independently using a specific output modality, leaving users to choose the preferred method of interaction in different scenarios. We evaluate the value of a shared input multimodal menu system in both a single-task desktop setting and in a dynamic dual-task setting, in which the user was required to interact with the shared input multimodal menu system while driving a simulated vehicle. Results indicate that users were faster at locating a target item in the menu when visual feedback was provided in the single-task desktop setting, but in the dual-task driving setting, visual output presented a significant source of visual distraction that interfered with driving performance. In contrast, auditory output mitigated some of the risk associated with menu selection while driving. A shared input multimodal interface allows users to take advantage of multiple feedback modalities properly, providing a better overall experience.

[1]  Ravin Balakrishnan,et al.  Simple vs. compound mark hierarchical marking menus , 2004, UIST '04.

[2]  Duncan P. Brumby,et al.  iPod distraction: effects of portable music-player use on driver performance , 2007, CHI.

[3]  Barry Arons,et al.  VoiceNotes: a speech interface for a hand-held voice notetaker , 1993, INTERCHI.

[4]  Stephen A. Brewster,et al.  Using nonspeech sounds to provide navigation cues , 1998, TCHI.

[5]  Meera Blattner,et al.  Earcons and Icons: Their Structure and Common Design Principles , 1989, Hum. Comput. Interact..

[6]  Dario D. Salvucci A Multitasking General Executive for Compound Continuous Tasks , 2005, Cogn. Sci..

[7]  Lisa J. Stifelman,et al.  Paper and Pen Interaction with Structured Speech , 2001 .

[8]  Paul Resnick,et al.  Skip and scan: cleaning up telephone interface , 1992, CHI '92.

[9]  Dwight P. Miller The Depth/Breadth Tradeoff in Hierarchical Computer Menus , 1981 .

[10]  Dinesh K. Pai,et al.  JASS: A JAVA AUDIO SYNTHESIS SYSTEM FOR PROGRAMMERS , 2001 .

[11]  Gordon Kurtenbach,et al.  The design and evaluation of marking menus , 1993 .

[12]  Angelika Salmen,et al.  DIALOG SYSTEMS IN THE TRAFFIC ENVIRONMENT , 2000 .

[13]  Susan J. Lederman,et al.  Design and Performance of a Tactile Shape Display Using RC Servomotors (Short Paper) , 2004 .

[14]  Stephen A. Brewster,et al.  Correcting menu usability problems with sound , 1999, Behav. Inf. Technol..

[15]  Gilles Bailly,et al.  Wave Menus: Improving the Novice Mode of Hierarchical Marking Menus , 2007, INTERACT.

[16]  George Engelbeck,et al.  Designing Voice Menu Applications for Telephones , 1997 .

[17]  Timothy L. Brown,et al.  Speech-Based Interaction with In-Vehicle Computers: The Effect of Speech-Based E-Mail on Drivers' Attention to the Roadway , 2001, Hum. Factors.

[18]  Ben Shneiderman,et al.  The Psychology of Menu Selection: Designing Cognitive Control at the Human/Computer Interface , 1991 .

[19]  A J McKnight,et al.  The effect of cellular phone use upon driver attention. , 1993, Accident; analysis and prevention.

[20]  Chris Schmandt,et al.  Nomadic radio: speech and audio interaction for contextual messaging in nomadic environments , 2000, TCHI.

[21]  Joëlle Coutaz,et al.  A design space for multimodal systems: concurrent processing and data fusion , 1993, INTERCHI.

[22]  Allen and Rosenbloom Paul S. Newell,et al.  Mechanisms of Skill Acquisition and the Law of Practice , 1993 .

[23]  Albrecht Schmidt,et al.  In-car interaction using search-based user interfaces , 2008, CHI.

[24]  Chunyuan Liao,et al.  Pen-top feedback for paper-based interfaces , 2006, UIST.

[25]  Christopher D. Wickens,et al.  Multiple resources and performance prediction , 2002 .

[26]  Stephen A. Brewster,et al.  Multimodal 'eyes-free' interaction techniques for wearable devices , 2003, CHI '03.

[27]  Sara B. Kiesler,et al.  Calling while driving: effects of providing remote traffic context , 2005, CHI.

[28]  David R. Morse,et al.  Using while moving: HCI issues in fieldwork environments , 2000, TCHI.

[29]  Barbara Freeman,et al.  Curing the menu blues in touch-tone voice interfaces , 2001, CHI Extended Abstracts.

[30]  Duncan P. Brumby,et al.  Fast or safe?: how performance objectives determine modality output choices while interacting on the move , 2011, CHI.

[31]  Julie A. Jacko,et al.  Older adults and visual impairment: what do exposure times and accuracy tell us about performance gains associated with multimodal feedback? , 2003, CHI '03.

[32]  Stephen A. Brewster,et al.  Maximising screen-space on mobile computing devices , 1999, CHI Extended Abstracts.

[33]  Ann Blandford,et al.  Four easy pieces for assessing the usability of multimodal interaction: the CARE properties , 1995, INTERACT.

[34]  David L. Strayer,et al.  Driven to Distraction: Dual-Task Studies of Simulated Driving and Conversing on a Cellular Telephone , 2001, Psychological science.

[35]  Stephen A. Brewster,et al.  Gestural and audio metaphors as a means of control for mobile devices , 2002, CHI.

[36]  J. Walker,et al.  In-vehicle navigation devices: Effects on the safety of driver performance , 1990, Vehicle Navigation and Information Systems Conference, 1991.

[37]  Mikael B. Skov,et al.  You can touch, but you can't look: interacting with in-vehicle systems , 2008, CHI.

[38]  Dario D. Salvucci Predicting the effects of in-car interface use on driver performance: an integrated model approach , 2001, Int. J. Hum. Comput. Stud..

[39]  Vincent Hayward,et al.  A role for haptics in mobile interaction: initial design using a handheld tactile display prototype , 2006, CHI.

[40]  Pierre Dragicevic,et al.  Earpod: eyes-free menu selection using touch input and reactive audio feedback , 2007, CHI.

[41]  Ben Shneiderman,et al.  Split menus: effectively using selection frequency to organize menus , 1994, TCHI.

[42]  William W. Gaver,et al.  AUDITORY ICONS IN LARGE-SCALE COLLABORATIVE ENVIRONMENTS , 1990, SGCH.

[43]  Shumin Zhai,et al.  The benefits of augmenting telephone voice menu navigation with visual browsing and search , 2006, CHI.

[44]  Barry Arons,et al.  SpeechSkimmer: a system for interactively skimming recorded speech , 1997, TCHI.

[45]  William Buxton,et al.  The limits of expert performance using hierarchic marking menus , 1993, INTERCHI.

[46]  William Buxton,et al.  User learning and performance with marking menus , 1994, CHI 1994.

[47]  Kristen L. Macuga,et al.  Predicting the effects of cellular-phone dialing on driver performance , 2002, Cognitive Systems Research.

[48]  Shumin Zhai,et al.  SHARK2: a large vocabulary shorthand writing system for pen-based computers , 2004, UIST '04.

[49]  Julie A. Jacko,et al.  Foundation for improved interaction by individuals with visual impairments through multimodal feedback , 2002, Universal Access in the Information Society.

[50]  James D. Foley,et al.  The human factors of computer graphics interaction techniques , 1984, IEEE Computer Graphics and Applications.

[51]  Chris Schmandt,et al.  NewsComm: a hand-held interface for interactive access to structured audio , 1996, CHI.

[52]  Cristy Ho,et al.  Multisensory In-Car Warning Signals for Collision Avoidance , 2007, Hum. Factors.

[53]  Mark S. Ackerman,et al.  Impromptu: managing networked audio applications for mobile users , 2004, MobiSys '04.

[54]  Teresa L. Roberts,et al.  The effects of device technology on the usability of advanced telephone functions , 1989, CHI '89.

[55]  Joonhwan Lee,et al.  Studying the effectiveness of MOVE: a contextually optimized in-vehicle navigation system , 2005, CHI.

[56]  D. Strayer,et al.  Cell phone-induced failures of visual attention during simulated driving. , 2003, Journal of experimental psychology. Applied.

[57]  Morten Fjeld,et al.  Exploring user motivations for eyes-free interaction on mobile devices , 2012, CHI.

[58]  Karel Brookhuis,et al.  The interaction between driving and in-vehicle information systems: comparison of results from laboratory, simulator and real-world studies , 2005 .

[59]  Stephen A. Brewster,et al.  An evaluation of earcons for use in auditory human-computer interfaces , 1993, INTERCHI.

[60]  H Alm,et al.  Changes in driver behaviour as a function of handsfree mobile phones--a simulator study. , 1994, Accident; analysis and prevention.

[61]  Gilles Bailly,et al.  Flower menus: a new type of marking menu with large menu breadth, within groups and efficient expert mode memorization , 2008, AVI '08.

[62]  Eric S. Lee,et al.  "The Psychology of Menu Selection: Designing Cognitive Control at the Human/Computer Interface, " by Kent L. Norman (Book Review) , 1993, Int. J. Man Mach. Stud..

[63]  William W. Gaver The SonicFinder: An Interface That Uses Auditory Icons , 1989, Hum. Comput. Interact..

[64]  Chris Schmandt,et al.  Dynamic Soundscape: mapping time to space for audio browsing , 1997, CHI.

[65]  Maneesh Agrawala,et al.  Zone and polygon menus: using relative position to increase the breadth of multi-stroke marking menus , 2006, CHI.

[66]  M Akamatsu,et al.  Please Scroll down for Article Ergonomics a Comparison of Tactile, Auditory, and Visual Feedback in a Pointing Task Using a Mouse-type Device , 2022 .

[67]  Chris Schmandt,et al.  Audio hallway: a virtual acoustic environment for browsing , 1998, UIST '98.

[68]  Andrew Howes,et al.  Focus on driving: how cognitive constraints shape the adaptation of strategy when dialing while driving , 2009, CHI.

[69]  Mark Billinghurst,et al.  A user study of auditory versus visual interfaces for use while driving , 2008, Int. J. Hum. Comput. Stud..

[70]  M. Weiser,et al.  An empirical comparison of pie vs. linear menus , 1988, CHI '88.

[71]  Bastian Pfleging,et al.  SpeeT: A Multimodal Interaction Style Combining Speech and Touch Interaction in Automotive Environments , 2011 .

[72]  Tilman Dingler,et al.  Learnabiltiy of Sound Cues for Environmental Features: Auditory Icons, Earcons, Spearcons, and Speech , 2008 .

[73]  Elizabeth D. Mynatt Transforming graphical interfaces into auditory interfaces , 1995, CHI 95 Conference Companion.

[74]  Bruce N. Walker,et al.  SPEARCONS: SPEECH-BASED EARCONS IMPROVE NAVIGATION PERFORMANCE IN AUDITORY MENUS , 2006 .