Sasayaki: augmented voice web browsing experience

Auditory user interfaces have great Web-access potential for billions of people with visual impairments, with limited literacy, who are driving, or who are otherwise unable to use a visual interface. However a sequential speech-based representation can only convey a limited amount of information. In addition, typical auditory user interfaces lose the visual cues such as text styles and page structures, and lack effective feedback about the current focus. To address these limitations, we created Sasayaki (from whisper in Japanese), which augments the primary voice output with a secondary whisper of contextually relevant information, automatically or in response to user requests. It also offers new ways to jump to semantically meaningful locations. A prototype was implemented as a plug-in for an auditory Web browser. Our experimental results show that the Sasayaki can reduce the task completion times for finding elements in webpages and increase satisfaction and confidence.

[1]  Frank Dellaert,et al.  SWAN: System for Wearable Audio Navigation , 2007, 2007 11th IEEE International Symposium on Wearable Computers.

[2]  Kitch Barnicle,et al.  Usability testing with screen reading technology in a Windows environment , 2000, CUU '00.

[3]  Hironobu Takagi,et al.  Analysis of navigability of Web applications for improving blind usability , 2007, TCHI.

[4]  Volker Scholz,et al.  An Augmented Reality museum guide , 2008, 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality.

[5]  Carole A. Goble,et al.  The travails of visually impaired web travellers , 2000, HYPERTEXT '00.

[6]  Hironobu Takagi,et al.  Accessibility commons: a metadata infrastructure for web accessibility , 2008, Assets '08.

[7]  I. V. Ramakrishnan,et al.  Csurf: a context-driven non-visual web-browser , 2007, WWW '07.

[8]  Matt Jones,et al.  ONTRACK: Dynamically adapting music playback to support navigation , 2008, Personal and Ubiquitous Computing.

[9]  Watanabe Hideo,et al.  Deeper Sentiment Analysis Using Machine Translation Technology , 2004, COLING.

[10]  Hironobu Takagi,et al.  Aibrowser for multimedia: introducing multimedia content accessibility for visually impaired users , 2007, Assets '07.

[11]  Max Mühlhäuser,et al.  AUGUR: providing context-aware interaction support , 2009, EICS '09.

[12]  Roy Carr-Hill,et al.  International literacy statistics: a review of concepts, methodology and current data , 2008 .

[13]  Hironobu Takagi,et al.  Social accessibility: achieving accessibility through collaborative metadata authoring , 2008, Assets '08.

[14]  David Salesin,et al.  Summarizing personal web browsing sessions , 2006, UIST.

[15]  Pattie Maes,et al.  Agents that reduce work and information overload , 1994, CACM.

[16]  Carole A. Goble,et al.  Evaluating DANTE: Semantic transcoding for visually disabled users , 2007, TCHI.

[17]  Benjamin B. Bederson,et al.  Audio augmented reality: a prototype automated tour guide , 1995, CHI 95 Conference Companion.

[18]  Peter Parente,et al.  Clique: a conversant, task-based audio display for GUI applications , 2006, ASAC.

[19]  Shaojian Zhu,et al.  Sasayaki: an augmented voice-based web browsing experience , 2010, ASSETS '10.

[20]  I. V. Ramakrishnan,et al.  What's new?: making web page updates accessible , 2008, Assets '08.

[21]  Brad A. Myers,et al.  Citrine: providing intelligent copy-and-paste , 2004, UIST '04.

[22]  Thierry Pun,et al.  Auditory browser for blind and visually impaired users , 1999, CHI Extended Abstracts.

[23]  Henry Lieberman,et al.  Supporting user hypotheses in problem diagnosis , 2004, IUI '04.

[24]  T. V. Raman,et al.  AxsJAX: a talking translation bot using google IM: bringing web-2.0 applications to life , 2008, W4A '08.

[25]  Gerhard Eckel,et al.  Immersive audio-augmented environments: the LISTEN project , 2001, Proceedings Fifth International Conference on Information Visualisation.

[26]  Pattie Maes,et al.  Agents that reduce work and information overload , 1994, CACM.

[27]  Aaron Allen,et al.  What Frustrates Screen Reader Users on the Web: A Study of 100 Blind Users , 2007, Int. J. Hum. Comput. Interact..

[28]  Marek Hatala,et al.  Using semantic web approach in augmented audio reality system for museum visitors , 2004, WWW Alt. '04.

[29]  Ravi Kuber,et al.  Improving web accessibility using content-aware plug-ins , 2005, CHI EA '05.

[30]  F. Williams Measuring the Information Society , 1988 .

[31]  Simon Harper,et al.  Gist summaries for visually impaired surfers , 2005, Assets '05.

[32]  Chris Schmandt,et al.  Nomadic radio: speech and audio interaction for contextual messaging in nomadic environments , 2000, TCHI.

[33]  J. Borenstein,et al.  The NavBelt-a computerized travel aid for the blind based on mobile robotics technology , 1998, IEEE Transactions on Biomedical Engineering.

[34]  Sean Bechhofer,et al.  The SADIe transcoding platform , 2008, W4A '08.