Modeling Expertise in Assistive Navigation Interfaces for Blind People

Evaluating the impact of expertise and route knowledge on task performance can guide the design of intelligent and adaptive navigation interfaces. Expertise has been relatively unexplored in the context of assistive indoor navigation interfaces for blind people. To quantify the complex relationship between the user»s walking patterns, route learning, and adaptation to the interface, we conducted a study with 8 blind participants. The participants repeated a set of navigation tasks while using a smartphone-based turn-by-turn navigation guidance app. The results demonstrate the gradual evolution of user skill and knowledge throughout the route repetitions, significantly impacting the task completion time. In addition to the exploratory analysis, we take a step towards tailoring the navigation interface to the user»s needs by proposing a personalized recurrent neural network-based behavior model for expertise level classification.

[1]  Anke M. Brock,et al.  Accessible Interactive Maps for Visually Impaired Users , 2022, ArXiv.

[2]  Chieko Asakawa,et al.  People with Visual Impairment Training Personal Object Recognizers: Feasibility and Challenges , 2017, CHI.

[3]  Rob Kitchin,et al.  The Effect of Spatial Tasks on Visually Impaired Peoples’ Wayfinding Abilities , 2002 .

[4]  Hironobu Takagi,et al.  NavCog3: An Evaluation of a Smartphone-Based Blind Indoor Navigation Assistant with Semantic Features in a Large-Scale Environment , 2017, ASSETS.

[5]  João Guerreiro,et al.  Virtual Navigation for Blind People: Building Sequential Representations of the Real-World , 2017, ASSETS.

[6]  Kostas E. Bekris,et al.  The user as a sensor: navigating users with visual impairments in indoor spaces using tactile landmarks , 2012, CHI.

[7]  Vladimir I. Levenshtein,et al.  Binary codes capable of correcting deletions, insertions, and reversals , 1965 .

[8]  Charles Elkan,et al.  Learning to Diagnose with LSTM Recurrent Neural Networks , 2015, ICLR.

[9]  Koji Yatani,et al.  SpaceSense: representing geographical information to visually impaired people using spatial tactile feedback , 2012, CHI.

[10]  Kostas E. Bekris,et al.  Indoor Human Navigation Systems: A Survey , 2013, Interact. Comput..

[11]  F. Drews,et al.  The Effects of Acoustic Turn-by-turn Navigation on Wayfinding , 2010 .

[12]  Andrew W. Senior,et al.  Long short-term memory recurrent neural network architectures for large scale acoustic modeling , 2014, INTERSPEECH.

[13]  Anind K. Dey,et al.  Uncovering information needs for independent spatial learning for users who are visually impaired , 2013, ASSETS.

[14]  Chieko Asakawa,et al.  Environmental Factors in Indoor Navigation Based on Real-World Trajectories of Blind Users , 2018, CHI.

[15]  Romedi Passini,et al.  The Spatio-Cognitive Abilities of the Visually Impaired Population , 1990 .

[16]  David Mioduser,et al.  Haptic-feedback support for cognitive mapping of unknown spaces by people who are blind , 2008, Int. J. Hum. Comput. Stud..

[17]  Scott Sanner,et al.  Deep Sequential Recommendation for Personalized Adaptive User Interfaces , 2017, IUI.

[18]  R. Kitchin,et al.  Techniques to Collect and Analyze the Cognitive Map Knowledge of Persons with Visual Impairment Or Blindness: Issues of Validity , 1997 .

[19]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[20]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[21]  Masayuki Murata,et al.  Achieving Practical and Accurate Indoor Navigation for People with Visual Impairments , 2017, W4A.

[22]  Hironobu Takagi,et al.  NavCog: a navigational cognitive assistant for the blind , 2016, MobileHCI.

[23]  Jaime Sánchez,et al.  Audio haptic videogaming for developing wayfinding skills in learners who are blind , 2014, IUI.

[24]  Fillia Makedon,et al.  CogniLearn: A Deep Learning-based Interface for Cognitive Behavior Assessment , 2017, IUI.