Automatic Detection of Urban Features from Wheelchair Users’ Movements

Providing city navigation instructions to people with motion disabilities requires the knowledge of urban features like curb ramps, steps or other obstacles along the way. Since these urban features are not available from maps and change in time, crowdsourcing this information from end-users is a scalable and promising solution. Our preliminary study on wheelchair users shows that an automatic crowdsourcing mechanism is needed, avoiding users’ involvement.In this contribution we present a solution to crowdsource urban features from inertial sensors installed on a wheelchair. Activity recognition techniques based on decision trees are used to process the sensors data stream. Experimental results, conducted with data acquired from 10 real wheelchair users navigating in an outdoor environment show that our solution is effective in detecting urban features with precision around 0.9, while it is less reliable when classifying some fine-grained urban feature characteristics, like a step height. The experimental results also present our investigation aimed at identifying the best parameters for the given problem, which include number, position and type of inertial sensors, classifier type, segmentation parameters, etc.

[1]  Francisco Herrera,et al.  SMOTE-RSB*: a hybrid preprocessing approach based on oversampling and undersampling for high imbalanced data-sets using SMOTE and rough sets theory , 2012, Knowledge and Information Systems.

[2]  Ming Ren,et al.  Movement Pattern Recognition Assisted Map Matching for Pedestrian/Wheelchair Navigation , 2012, Journal of Navigation.

[3]  Alex A. Freitas,et al.  A survey of hierarchical classification across different application domains , 2010, Data Mining and Knowledge Discovery.

[4]  Jon Froehlich,et al.  A Pilot Deployment of an Online Tool for Large-Scale Virtual Auditing of Urban Accessibility , 2017, ASSETS.

[5]  Michael Kipp,et al.  ANVIL - a generic annotation tool for multimodal dialogue , 2001, INTERSPEECH.

[6]  Roberto Manduchi,et al.  Mind Your Crossings , 2017, ACM Trans. Access. Comput..

[7]  Jyri Rajamäki,et al.  LaureaPOP indoor navigation service for the visually impaired in a WLAN environment , 2007 .

[8]  Raquel Velho Transport accessibility for wheelchair users: A qualitative analysis of inclusion and health , 2019, International Journal of Transportation Science and Technology.

[9]  Gerhard Weber,et al.  RouteCheckr: personalized multicriteria routing for mobility impaired pedestrians , 2008, Assets '08.

[10]  Paul J. M. Havinga,et al.  A Survey of Online Activity Recognition Using Mobile Phones , 2015, Sensors.

[11]  Dragan Ahmetovic,et al.  ZebraRecognizer: Pedestrian crossing recognition for people with visual impairment or blindness , 2016, Pattern Recognit..

[12]  Zhan Liu,et al.  How to motivate participation and improve quality of crowdsourcing when building accessibility maps , 2018, 2018 15th IEEE Annual Consumer Communications & Networking Conference (CCNC).

[13]  Alessandro Rizzi,et al.  Robust traffic lights detection on mobile devices for pedestrians with visual impairment , 2016, Comput. Vis. Image Underst..

[14]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[15]  Hung Manh La,et al.  A Comprehensive Review of Smart Wheelchairs: Past, Present, and Future , 2017, IEEE Transactions on Human-Machine Systems.

[16]  Bernt Schiele,et al.  A tutorial on human activity recognition using body-worn inertial sensors , 2014, CSUR.

[17]  Gary M. Weiss,et al.  Smartwatch-based activity recognition: A machine learning approach , 2016, 2016 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI).

[18]  H. Matthews,et al.  Mapping for Wheelchair Users: Route Navigation in Urban Spaces , 2006 .

[19]  Shinichiro Haruyama,et al.  Indoor navigation system for visually impaired people using visible light communication and compensated geomagnetic sensing , 2012, 2012 1st IEEE International Conference on Communications in China (ICCC).

[20]  Masayuki Murata,et al.  Smartphone-based Indoor Localization for Blind Navigation across Building Complexes , 2018, 2018 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[21]  Roberto Bresin,et al.  A Systematic Review of Mapping Strategies for the Sonification of Physical Quantities , 2013, PloS one.

[22]  Claudio E. Palazzi,et al.  Movement pattern recognition through smartphone's accelerometer , 2012, 2012 IEEE Consumer Communications and Networking Conference (CCNC).

[23]  Juan-Luis Gorricho,et al.  Activity Recognition from Accelerometer Data on a Mobile Phone , 2009, IWANN.

[24]  Giancarlo Fortino,et al.  Activity Level Assessment Using a Smart Cushion for People with a Sedentary Lifestyle , 2017, Sensors.

[25]  Chieko Asakawa,et al.  Turn Right: Analysis of Rotation Errors in Turn-by-Turn Navigation for Individuals with Visual Impairments , 2018, ASSETS.