Use of Speech Technology in Real Life Environment

This paper reports results from two six-month field studies investigating the use of speech-based interactions in real world environments. The first study focused on the use of speech-based dictation/word processing tasks that allow users to generate text such as reports or emails. The second study took a broader view and examined two types of interactions: speech-based dictation for word processing and speech-based command and control supporting interactions with the graphical user interface. The results suggest that user requirements for speech-based interactions have evolved as the technology matured to include better support for formatting text documents as well as more effective support for other applications. While significant research has focused on the use of speech recognition to generate text, our participants spent substantially more time using speech recognition for other, often unexpected tasks such as instant messaging and web browsing. The use of speech recognition to support informal communication is particularly interesting as accuracy may be less critical in this context. Further research is needed to address these emerging requirements for speech technologies.

[1]  K. Hux,et al.  Speech recognition training for enhancing written language generation by a traumatic brain injury survivor. , 2000, Brain injury.

[2]  Shaojian Zhu,et al.  Speech-Based Navigation: Improving Grid-Based Solutions , 2009, INTERACT.

[3]  Clare-Marie Karat,et al.  The Beauty of Errors: Patterns of Error Correction in Desktop Speech Systems , 1999, INTERACT.

[4]  Clare-Marie Karat,et al.  Hands-Free, Speech-Based Navigation During Dictation: Difficulties, Consequences, and Solutions , 2003, Hum. Comput. Interact..

[5]  Pavel Slavík,et al.  Non-speech input and speech recognition for real-time control of computer games , 2006, Assets '06.

[6]  Bill Z. Manaris,et al.  SUITEKeys: a speech understanding interface for the motor-control challenged , 1998, Assets '98.

[7]  Ben Shneiderman,et al.  A comparison of voice controlled and mouse controlled web browsing , 2000, Assets '00.

[8]  Andrew Sears,et al.  Speech-based cursor control , 2002, ASSETS.

[9]  James A. Landay,et al.  Voicedraw: a hands-free voice-driven drawing application for people with motor impairments , 2007, Assets '07.

[10]  Etsuya Shibayama,et al.  The migratory cursor: accurate speech-based cursor movement by moving multiple ghost cursors using non-verbal vocalizations , 2005, Assets '05.

[11]  Chris Baber,et al.  Modelling Error Recovery and Repair in Automatic Speech Recognition , 1993, Int. J. Man Mach. Stud..

[12]  Clare-Marie Karat,et al.  Conversational interface technologies , 2002 .

[13]  Human-Computer Interaction – INTERACT 2009 , 2009, Lecture Notes in Computer Science.

[14]  Clare-Marie Karat,et al.  How productivity improves in hands-free continuous dictation tasks: lessons learned from a longitudinal study , 2005, Interact. Comput..

[15]  Xiao Li,et al.  The vocal joystick:: evaluation of voice-based cursor control techniques , 2006, Assets '06.

[16]  Andrew Sears,et al.  Speech-based cursor control using grids: modelling performance and comparisons with other solutions , 2005, Behav. Inf. Technol..

[17]  Clare-Marie Karat,et al.  Productivity, satisfaction, and interaction strategies of individuals with spinal cord injuries and traditional users interacting with speech recognition software , 2001, Universal Access in the Information Society.

[18]  David D. Langan,et al.  A voice-activated syntax-directed editor for manually disabled programmers , 2006, Assets '06.

[19]  James A. Landay,et al.  Sketching images eyes-free: a grid-based dynamic drawing tool for the blind , 2002, Assets '02.

[20]  Stephen C. Arnold,et al.  Programming by voice, VocalProgramming , 2000, Assets '00.

[21]  Alfred Kobsa,et al.  Talking to the ceiling: an interface for bed-ridden manually impaired users , 1999, CHI EA '99.

[22]  Andrew Sears,et al.  Using confidence scores to improve hands-free speech based navigation in continuous dictation systems , 2004, TCHI.

[23]  Alexander H. Waibel,et al.  Improving recognizer acceptance through robust, natural speech repair , 1994, ICSLP.