Ubiquitous Text Interaction

Computer-based interactions increasingly pervade our everyday environments. Be it on a mobile device, a wearable device, a wall-sized display, or an augmented reality device, interactive systems often rely on the consumption, composition, and manipulation of text. The focus of this workshop is on exploring the problems and opportunities of text interactions that are embedded in our environments, available all the time, and used by people who may be constrained by device, situation, or disability. This workshop welcomes all researchers interested in interactive systems that rely on text input or output. Participants should submit a short position statement outlining their background, past work, future plans, and suggesting a use-case they would like to explore in-depth during the workshop. During the workshop, small teams will form around common or compelling use-cases. Teams will spend time brainstorming, creating low-fidelity prototypes, and discussing their use-case with the group. Participants may optionally submit a technical paper for presentation as part of the workshop program. The workshop serves to sustain and build the community of text entry researchers who attend CHI. It provides an opportunity for new members to join this community, soliciting feedback from experts in a small and supportive environment.

[1]  Mark D. Dunlop,et al.  Designing and evaluating text entry methods , 2012, CHI EA '12.

[2]  Mark D. Dunlop,et al.  Inviscid Text Entry and Beyond , 2016, CHI Extended Abstracts.

[3]  Mark D. Dunlop,et al.  Designed with Older Adults to SupportBetter Error Correction in SmartPhone Text Entry: The MaxieKeyboard , 2015, MobileHCI Adjunct.

[4]  Jacob O. Wobbrock,et al.  WalkType: using accelerometer data to accomodate situational impairments in mobile touch screen text entry , 2012, CHI.

[5]  Geehyuk Lee,et al.  Evaluation of a Smart-Restorable Backspace Technique to Facilitate Text Entry Error Correction , 2016, CHI.

[6]  Mark D. Dunlop,et al.  Text Entry on the Edge , 2015, CHI Extended Abstracts.

[7]  Woontack Woo,et al.  ARKB: 3D vision-based Augmented Reality Keyboard , 2003, ICAT.

[8]  Richard E. Ladner,et al.  Input finger detection for nonvisual touch screen text entry in Perkinput , 2012, Graphics Interface.

[9]  Gregory D. Abowd,et al.  An evaluation of BrailleTouch: mobile touchscreen text entry for the visually impaired , 2012, Mobile HCI.

[10]  Doug A. Bowman,et al.  Text Input Techniques for Immersive Virtual Environments: An Empirical Comparison , 2002 .

[11]  Luís Carriço,et al.  Effect of target size on non-visual text-entry , 2016, MobileHCI.

[12]  Shumin Zhai,et al.  Effects of Language Modeling and its Personalization on Touchscreen Typing Performance , 2015, CHI.

[13]  Torsten Kuhlen,et al.  SWIFTER: Design and evaluation of a speech-based text input metaphor for immersive virtual environments , 2016, 2016 IEEE Symposium on 3D User Interfaces (3DUI).

[14]  H Emslie,et al.  Text entry on handheld computers by older users , 2000, Ergonomics.

[15]  Brad A. Myers,et al.  EdgeWrite: a stylus-based text entry method designed for high accuracy and stability of motion , 2003, UIST '03.

[16]  Janet C. Read,et al.  Oops! silly me! errors in a handwriting recognition-based text entry interface for children , 2002, NordiCHI '02.

[17]  Roderick Murray-Smith,et al.  A Dose of Reality: Overcoming Usability Challenges in VR Head-Mounted Displays , 2015, CHI Extended Abstracts.

[18]  Ali Mazalek,et al.  Learning New Words and Spelling with Autocorrections , 2016, ISS.

[19]  Benjamin B. Bederson,et al.  Multi-modal text entry and selection on a mobile device , 2010, Graphics Interface.

[20]  Per Ola Kristensson,et al.  Parakeet: a continuous speech recognition system for mobile touch-screen devices , 2009, IUI.

[21]  Mark D. Dunlop,et al.  Grand challenges in text entry , 2013, CHI Extended Abstracts.

[22]  Mark D. Dunlop,et al.  The usability of text entry systems now and in the future , 2014, CHI Extended Abstracts.

[23]  Kyle Montague,et al.  B#: chord-based correction for multitouch braille input , 2014, CHI.

[24]  Luís Carriço,et al.  UbiBraille: designing and evaluating a vibrotactile Braille-reading device , 2013, ASSETS.

[25]  Shumeet Baluja,et al.  The role of context in query input: using contextual signals to complete queries on mobile devices , 2007, Mobile HCI.

[26]  Kyle Montague,et al.  Typing Performance of Blind Users: An Analysis of Touch Behaviors, Learning Effect, and In-Situ Usage , 2015, ASSETS.

[27]  Vittorio Fuccella,et al.  Gestures and widgets: performance in text editing on multi-touch capable mobile devices , 2013, CHI.