NavCog3: An Evaluation of a Smartphone-Based Blind Indoor Navigation Assistant with Semantic Features in a Large-Scale Environment

Navigating in unfamiliar environments is challenging for most people, especially for individuals with visual impairments. While many personal navigation tools have been proposed to enable in- dependent indoor navigation, they have insufficient accuracy (e.g., 5-10 m), do not provide semantic features about surroundings (e.g., doorways, shops, etc.), and may require specialized devices to function. Moreover, the deployment of many systems is often only evaluated in constrained scenarios, which may not precisely reflect the performance in the real world. Therefore, we have de- signed and implemented NavCog3, a smartphone-based indoor navigation assistant that has been evaluated in a 21,000 m2 shop- ping mall. In addition to turn-by-turn instructions, it provides in- formation on landmarks (e.g., tactile paving) and points of interests nearby. We first conducted a controlled study with 10 visually im- paired users to assess localization accuracy and the perceived use- fulness of semantic features. To understand the usability of the app in a real-world setting, we then conducted another study with 43 participants with visual impairments where they could freely nav- igate in the shopping mall using NavCog3. Our findings suggest that NavCog3 can open a new opportunity for users with visual im- pairments to independently find and visit large and complex places with confidence.

[1]  John Krumm,et al.  Accuracy characterization for metropolitan-scale Wi-Fi localization , 2005, MobiSys '05.

[2]  Sri Hastuti Kurniawan,et al.  A qualitative study to support a blind photography mobile application , 2013, PETRA '13.

[3]  V. Braun,et al.  Using thematic analysis in psychology , 2006 .

[4]  Uran Oh,et al.  Current and future mobile and wearable device use by people with visual impairments , 2014, CHI.

[5]  Ken Sakamura,et al.  Navigating visually impaired travelers in a large train station using smartphone and bluetooth low energy , 2016, SAC.

[6]  Dongsoo Han,et al.  Performance Evaluation of Radio Map Construction Methods for Wi-Fi Positioning Systems , 2017, IEEE Transactions on Intelligent Transportation Systems.

[7]  Abdelsalam Helal,et al.  Drishti: an integrated indoor/outdoor blind navigation system and service , 2004, Second IEEE Annual Conference on Pervasive Computing and Communications, 2004. Proceedings of the.

[8]  Roberto Manduchi,et al.  (Computer) vision without sight , 2012, Commun. ACM.

[9]  Uran Oh,et al.  Design of and subjective response to on-body input for people with visual impairments , 2014, ASSETS.

[10]  Amy Hurst,et al.  "Pray before you step out": describing personal and situational blind navigation behaviors , 2013, ASSETS.

[11]  Ramiro Velazquez,et al.  Wearable Assistive Devices for the Blind , 2016, ArXiv.

[12]  Roberto Manduchi,et al.  The last meter: blind visual guidance to a target , 2014, CHI.

[13]  Ravi Kuber,et al.  An empirical investigation of the situationally-induced impairments experienced by blind mobile device users , 2016, W4A.

[14]  Jayakorn Vongkulbhisal,et al.  Beacon-Guided Structure from Motion for Smartphone-Based Navigation , 2017, 2017 IEEE Winter Conference on Applications of Computer Vision (WACV).

[15]  Shiri Azenkot,et al.  Exploring the use of speech input by blind people on mobile devices , 2013, ASSETS.

[16]  Masayuki Murata,et al.  Achieving Practical and Accurate Indoor Navigation for People with Visual Impairments , 2017, W4A.

[17]  Hironobu Takagi,et al.  Assessment of Semantic Taxonomies for Blind Indoor Navigation Based on a Shopping Center Use Case , 2017, W4A.

[18]  Karen L. Myers,et al.  Task Management under Change and Uncertainty Constraint Solving Experience with the CALO Project , 2005 .

[19]  Chieko Asakawa,et al.  LuzDeploy: A Collective Action System for Installing Navigation Infrastructure for Blind People , 2017, W4A.

[20]  Lionel M. Ni,et al.  A Survey on Wireless Indoor Localization from the Device Perspective , 2016, ACM Comput. Surv..

[21]  Thomas Ertl,et al.  Design and development of an indoor navigation and object identification system for the blind , 2003, ASSETS.

[22]  M Dias Indoor Navigation Challenges for Visually Impaired People , 2015 .

[23]  Kostas E. Bekris,et al.  Indoor Human Navigation Systems: A Survey , 2013, Interact. Comput..

[24]  Gordon E. Legge,et al.  Blind Navigation and the Role of Technology , 2008 .

[25]  J Faria,et al.  Electronic white cane for blind people navigation assistance , 2010, 2010 World Automation Congress.

[26]  Roberto Manduchi,et al.  Blind guidance using mobile computer vision: a usability study , 2010, ASSETS '10.

[27]  Marilyn Tremaine CHI '01 Extended Abstracts on Human Factors in Computing Systems , 2001, CHI Extended Abstracts.

[28]  Richard E. Ladner,et al.  Freedom to roam: a study of mobile device adoption and accessibility for people with visual and motor disabilities , 2009, Assets '09.

[29]  J. Lovie-Kitchin,et al.  Mobility performance with retinitis pigmentosa , 1997 .

[30]  Christophe Jouffrais,et al.  Route selection algorithm for Blind pedestrian , 2010, ICCAS 2010.

[31]  Hironobu Takagi,et al.  NavCog: turn-by-turn smartphone navigation assistant for people with visual impairments or blindness , 2016, W4A.

[32]  Dierna Giovanni Luca Towards accurate indoor localization using iBeacons , fingerprinting and particle filtering , 2016 .

[33]  David A. Ross,et al.  Implementing Assistive Technology on Wearable Computers , 2001, IEEE Intell. Syst..

[34]  Mark W. Newman,et al.  Supporting visually impaired navigation: a needs-finding study , 2011, CHI Extended Abstracts.

[35]  Kostas E. Bekris,et al.  The user as a sensor: navigating users with visual impairments in indoor spaces using tactile landmarks , 2012, CHI.

[36]  William C. Mann,et al.  Smart Phone Based Cognitive Assistant , 2003 .

[37]  Sumi Helal,et al.  The Engineering Handbook of Smart Technology for Aging, Disability, and Independence , 2008 .