Insights on Assistive Orientation and Mobility of People with Visual Impairment Based on Large-Scale Longitudinal Data

Assistive applications for orientation and mobility promote independence for people with visual impairment (PVI). While typical design and evaluation of such applications involves small-sample iterative studies, we analyze large-scale longitudinal data from a geographically diverse population. Our publicly released dataset from iMove, a mobile app supporting orientation of PVI, contains millions of interactions by thousands of users over a year. Our analysis (i) examines common functionalities, settings, assistive features, and movement modalities in iMove dataset and (ii) discovers user communities based on interaction patterns. We find that the most popular interaction mode is passive, where users receive more notifications, often verbose, while in motion and perform fewer actions. The use of built-in assistive features such as enlarged text indicate a high presence of users with residual sight. Users fall into three distinct groups: (C1) users interested in surrounding points of interest, (C2) users interacting in short bursts to inquire about current location, and (C3) users with long active sessions while in motion. iMove was designed with C3 in mind, and one strength of our contribution is providing meaningful semantics for unanticipated groups, C1 and C2. Our analysis reveals insights that can be generalized to other assistive orientation and mobility applications.

[1]  Alessandro Rizzi,et al.  Robust traffic lights detection on mobile devices for pedestrians with visual impairment , 2016, Comput. Vis. Image Underst..

[2]  Hironobu Takagi,et al.  NavCog: turn-by-turn smartphone navigation assistant for people with visual impairments or blindness , 2016, W4A.

[3]  A. Azzouz 2011 , 2020, City.

[4]  Dragan Ahmetovic,et al.  ZebraRecognizer: Efficient and Precise Localization of Pedestrian Crossings , 2014, 2014 22nd International Conference on Pattern Recognition.

[5]  Anil K. Jain Data clustering: 50 years beyond K-means , 2008, Pattern Recognit. Lett..

[6]  Gian Paolo Rossi,et al.  Clique-aware mobile social clouds , 2016, 2016 IFIP Networking Conference (IFIP Networking) and Workshops.

[7]  Dragan Ahmetovic,et al.  ZebraRecognizer: Pedestrian crossing recognition for people with visual impairment or blindness , 2016, Pattern Recognit..

[8]  MascettiSergio,et al.  Sonification of guidance data during road crossing for people with visual impairments or blindness , 2016 .

[9]  Shinichiro Haruyama,et al.  New indoor navigation system for visually impaired people using visible light communication , 2013, EURASIP J. Wirel. Commun. Netw..

[10]  Chieko Asakawa,et al.  People with Visual Impairment Training Personal Object Recognizers: Feasibility and Challenges , 2017, CHI.

[11]  MascettiSergio,et al.  Robust traffic lights detection on mobile devices for pedestrians with visual impairment , 2016 .

[12]  Zhicheng Liu,et al.  Identifying Frequent User Tasks from Application Logs , 2017, IUI.

[13]  Scott E. Hudson,et al.  Distinguishing Users By Pointing Performance in Laboratory and Real-World Tasks , 2013, TACC.

[14]  Johannes Schöning,et al.  Falling asleep with Angry Birds, Facebook and Kindle: a large scale study on mobile application usage , 2011, Mobile HCI.

[15]  C. Thinus-Blanc,et al.  Representation of space in blind persons: vision as a spatial sense? , 1997, Psychological bulletin.

[16]  M Damashek,et al.  Gauging Similarity with n-Grams: Language-Independent Categorization of Text , 1995, Science.

[17]  Yong Jeong,et al.  Spatial cognition , 2004, Neurology.

[18]  João Guerreiro,et al.  Virtual Navigation for Blind People: Building Sequential Representations of the Real-World , 2017, ASSETS.

[19]  Amy Hurst,et al.  "Pray before you step out": describing personal and situational blind navigation behaviors , 2013, ASSETS.

[20]  Richard E. Ladner,et al.  WebinSitu: a comparative analysis of blind and sighted browsing behavior , 2007, Assets '07.

[21]  Kumar Yelamarthi,et al.  RFID and GPS integrated navigation system for the visually impaired , 2010, 2010 53rd IEEE International Midwest Symposium on Circuits and Systems.

[22]  Stephen M. Kosslyn,et al.  Mental imagery and sensory experience in congenital blindness , 1988, Neuropsychologia.

[23]  Huan Liu,et al.  Community Detection and Mining in Social Media , 2010, Community Detection and Mining in Social Media.

[24]  Kostas E. Bekris,et al.  Indoor Human Navigation Systems: A Survey , 2013, Interact. Comput..

[25]  Hironobu Takagi,et al.  Exploring Interface Design for Independent Navigation by People with Visual Impairments , 2015, ASSETS.

[26]  Claudio Bettini,et al.  Fine-grained recognition of abnormal behaviors for early detection of mild cognitive impairment , 2015, 2015 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[27]  Abdelsalam Helal,et al.  Drishti: an integrated indoor/outdoor blind navigation system and service , 2004, Second IEEE Annual Conference on Pervasive Computing and Communications, 2004. Proceedings of the.

[28]  J. G. Adair,et al.  The Hawthorne effect: A reconsideration of the methodological artifact. , 1984 .

[29]  Hironobu Takagi,et al.  NavCog: a navigational cognitive assistant for the blind , 2016, MobileHCI.

[30]  Claudio Bettini,et al.  SmartFABER: Recognizing fine-grained abnormal behaviors for early detection of mild cognitive impairment , 2016, Artif. Intell. Medicine.

[31]  Yiming Yang,et al.  A Comparative Study on Feature Selection in Text Categorization , 1997, ICML.

[32]  M E J Newman,et al.  Finding and evaluating community structure in networks. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.

[33]  Navneet Kaur,et al.  Opinion mining and sentiment analysis , 2016, 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom).

[34]  Torben Poulsen Acoustic traffic signal for blind pedestrians , 1982 .

[35]  Sushil Jajodia,et al.  Anonymity in Location-Based Services: Towards a General Framework , 2007, 2007 International Conference on Mobile Data Management.

[36]  Tyler Thrash,et al.  Spatial navigation by congenitally blind individuals , 2015, Wiley interdisciplinary reviews. Cognitive science.

[37]  Richard A. Harshman,et al.  Indexing by Latent Semantic Analysis , 1990, J. Am. Soc. Inf. Sci..

[38]  Daniel Gatica-Perez,et al.  Smartphone usage in the wild: a large-scale analysis of applications and context , 2011, ICMI '11.

[39]  Dragan Ahmetovic,et al.  Sonification of guidance data during road crossing for people with visual impairments or blindness , 2015, Int. J. Hum. Comput. Stud..

[40]  Xing Xie,et al.  Mining user similarity based on location history , 2008, GIS '08.

[41]  David Elsweiler,et al.  Using Sessions from Clickstream Data Analysis to Uncover Different Types of Twitter Behaviour , 2017, ISI.

[42]  R. Welsh Foundations of Orientation and Mobility , 1979 .

[43]  J. MacQueen Some methods for classification and analysis of multivariate observations , 1967 .

[44]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[45]  N. Smelser,et al.  International Encyclopedia of the Social and Behavioral Sciences , 2001 .

[46]  Gang Wang,et al.  Unsupervised Clickstream Clustering for User Behavior Analysis , 2016, CHI.

[47]  Hironobu Takagi,et al.  Supporting Orientation of People with Visual Impairment: Analysis of Large Scale Usage Data , 2016, ASSETS.

[48]  Masayuki Murata,et al.  Achieving Practical and Accurate Indoor Navigation for People with Visual Impairments , 2017, W4A.

[49]  Jean-Loup Guillaume,et al.  Fast unfolding of communities in large networks , 2008, 0803.0476.

[50]  Dragan Ahmetovic,et al.  Zebralocalizer: identification and localization of pedestrian crossings , 2011, Mobile HCI.

[51]  Gordon E. Legge,et al.  Blind Navigation and the Role of Technology , 2008 .