Wearable Travel Aid for Environment Perception and Navigation of Visually Impaired People

Assistive devices for visually impaired people (VIP) which support daily traveling and improve social inclusion are developing fast. Most of them try to solve the problem of navigation or obstacle avoidance, and other works focus on helping VIP to recognize their surrounding objects. However, very few of them couple both capabilities (i.e., navigation and recognition). Aiming at the above needs, this paper presents a wearable assistive device that allows VIP to (i) navigate safely and quickly in unfamiliar environment, and (ii) to recognize the objects in both indoor and outdoor environments. The device consists of a consumer Red, Green, Blue and Depth (RGB-D) camera and an Inertial Measurement Unit (IMU), which are mounted on a pair of eyeglasses, and a smartphone. The device leverages the ground height continuity among adjacent image frames to segment the ground accurately and rapidly, and then search the moving direction according to the ground. A lightweight Convolutional Neural Network (CNN)-based object recognition system is developed and deployed on the smartphone to increase the perception ability of VIP and promote the navigation system. It can provide the semantic information of surroundings, such as the categories, locations, and orientations of objects. Human–machine interaction is performed through audio module (a beeping sound for obstacle alert, speech recognition for understanding the user commands, and speech synthesis for expressing semantic information of surroundings). We evaluated the performance of the proposed system through many experiments conducted in both indoor and outdoor scenarios, demonstrating the efficiency and safety of the proposed assistive system.

[1]  Mark Sandler,et al.  MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[2]  Mun-Cheon Kang,et al.  An enhanced obstacle avoidance method for the visually impaired using deformable grid , 2017, IEEE Transactions on Consumer Electronics.

[3]  Farid Melgani,et al.  Indoor object recognition in RGBD images with complex-valued neural networks for visually-impaired people , 2019, Neurocomputing.

[4]  Teng Wang,et al.  A new method for obstacle detection based on Kinect depth image , 2015, 2015 Chinese Automation Congress (CAC).

[5]  Ruxandra Tapu,et al.  DEEP-SEE: Joint Object Detection, Tracking and Recognition with Application to Visually Impaired Navigational Assistance , 2017, Sensors.

[6]  Mun-Cheon Kang,et al.  A novel obstacle detection method based on deformable grid for the visually impaired , 2015, IEEE Transactions on Consumer Electronics.

[7]  Shiguo Lian,et al.  Smart guiding glasses for visually impaired people in indoor environment , 2017, IEEE Transactions on Consumer Electronics.

[8]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[9]  Xiangyu Zhang,et al.  ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[10]  Emanuele Frontoni,et al.  Mechatronic System to Help Visually Impaired Users During Walking and Running , 2018, IEEE Transactions on Intelligent Transportation Systems.

[11]  N. Otsu A threshold selection method from gray level histograms , 1979 .

[12]  António Pereira,et al.  Blind Guide: Anytime, Anywhere Solution for Guiding Blind People , 2017, WorldCIST.

[13]  A. Arditi,et al.  User Interface Preferences in the Design of a Camera-Based Navigation and Wayfinding Aid , 2013 .

[14]  Josechu J. Guerrero,et al.  Navigation Assistance for the Visually Impaired Using RGB-D Sensor With Range Expansion , 2016, IEEE Systems Journal.

[15]  Jian Bai,et al.  Detecting Traversable Area and Water Hazards for the Visually Impaired with a pRGB-D Sensor , 2017, Sensors.

[16]  Franco Penizzotto,et al.  Laser Radar Based Autonomous Mobile Robot Guidance System for Olive Groves Navigation , 2015, IEEE Latin America Transactions.

[17]  Yingli Tian,et al.  RGB-D Sensor-Based Computer Vision Assistive Technology for Visually Impaired Persons , 2014 .

[18]  Ruxandra Tapu,et al.  A computer vision system that ensure the autonomous navigation of blind people , 2013, 2013 E-Health and Bioengineering Conference (EHB).

[19]  Gopu Darsan,et al.  Computer Vision based Assistive Technology for Blind and Visually Impaired People , 2016, ICCCNT.

[20]  Jian Bai,et al.  Assisting the visually impaired: multitarget warning through millimeter wave radar and RGB-depth sensors , 2019, J. Electronic Imaging.

[21]  Bing Li,et al.  Vision-Based Mobile Indoor Assistive Navigation Aid for Blind People , 2019, IEEE Transactions on Mobile Computing.

[22]  Alexandru Gacsádi,et al.  Smartphone application to assist visually impaired people , 2017, 2017 14th International Conference on Engineering of Modern Electric Systems (EMES).

[23]  Luis Miguel Bergasa,et al.  Unifying Terrain Awareness for the Visually Impaired through Real-Time Semantic Segmentation , 2018, Sensors.

[24]  Cang Ye,et al.  Co-Robotic Cane: A New Robotic Navigation Aid for the Visually Impaired , 2016, IEEE Systems, Man, and Cybernetics Magazine.

[25]  Jinqiang Bai,et al.  Virtual-Blind-Road Following-Based Wearable Navigation Device for Blind People , 2018, IEEE Transactions on Consumer Electronics.

[26]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Zsolt T. Kardkovacs,et al.  Blind guide - A virtual eye for guiding indoor and outdoor movement , 2014, 2014 5th IEEE Conference on Cognitive Infocommunications (CogInfoCom).

[28]  Cang Ye,et al.  3-D Object Recognition of a Robotic Navigation Aid for the Visually Impaired , 2018, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[29]  Liang-Gee Chen,et al.  An intelligent depth-based obstacle detection system for visually-impaired aid applications , 2012, 2012 13th International Workshop on Image Analysis for Multimedia Interactive Services.

[30]  Denis Pomorski,et al.  GPS/IMU data fusion using multisensor Kalman filtering: introduction of contextual aspects , 2006, Inf. Fusion.

[31]  Juan D. Tardós,et al.  ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras , 2016, IEEE Transactions on Robotics.

[32]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[33]  Safvan Vahora,et al.  Android Smartphone Based Visual Object Recognition for Visually Impaired Using Deep Learning , 2018, 2018 International Conference on Communication and Signal Processing (ICCSP).

[34]  Andrew Blake,et al.  Multiscale Categorical Object Recognition Using Contour Fragments , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[35]  Rainer Stiefelhagen,et al.  Accessible section detection for visual guidance , 2013, 2013 IEEE International Conference on Multimedia and Expo Workshops (ICMEW).

[36]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.

[37]  Mohamed Lamine Mekhalfi,et al.  Recovering the Sight to blind People in indoor Environments with smart Technologies , 2016, Expert Syst. Appl..

[38]  Takeshi Saitoh,et al.  Indoor Mobile Robot Navigation by Central Following Based on Monocular Vision , 2009 .

[39]  Rabia Jafri,et al.  Computer vision-based object recognition for the visually impaired in an indoors environment: a survey , 2013, The Visual Computer.

[40]  Nikolaos Bourbakis,et al.  A System-Prototype Representing 3D Space via Alternative-Sensing for Visually Impaired Navigation , 2013, IEEE Sensors Journal.

[41]  Manjunatha Mahadevappa,et al.  Design, Development, and Clinical Evaluation of the Electronic Mobility Cane for Vision Rehabilitation , 2014, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[42]  Pei-Ying Chiang,et al.  Simple Smartphone-Based Guiding System for Visually Impaired People , 2017, Sensors.

[43]  Nikolaos G. Bourbakis,et al.  Wearable Obstacle Avoidance Electronic Travel Aids for Blind: A Survey , 2010, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[44]  Maamar Bettayeb,et al.  A Navigation Aid for Blind People , 2011, J. Intell. Robotic Syst..

[45]  François Chollet,et al.  Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[46]  Christoph M. Friedrich,et al.  Object Detection Featuring 3D Audio Localization for Microsoft HoloLens - A Deep Learning based Sensor Substitution Approach for the Blind , 2018, HEALTHINF.

[47]  Andrei Bursuc,et al.  A Smartphone-Based Obstacle Detection and Classification System for Assisting Visually Impaired People , 2013, 2013 IEEE International Conference on Computer Vision Workshops.

[48]  Mica R. Endsley,et al.  Toward a Theory of Situation Awareness in Dynamic Systems , 1995, Hum. Factors.

[49]  G. Reina,et al.  A new approach for terrain analysis in mobile robot applications , 2013, 2013 IEEE International Conference on Mechatronics (ICM).

[50]  Charles X. Ling,et al.  Pelee: A Real-Time Object Detection System on Mobile Devices , 2018, NeurIPS.

[51]  Erfu Yang,et al.  Review of machine vision-based electronic travel aids , 2017, 2017 23rd International Conference on Automation and Computing (ICAC).

[52]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[53]  Md. Milon Islam,et al.  Developing Walking Assistants for Visually Impaired People: A Review , 2019, IEEE Sensors Journal.

[54]  Luis Miguel Bergasa,et al.  Assisting the Visually Impaired: Obstacle Detection and Warning System by Acoustic Feedback , 2012, Sensors.

[55]  Jian Bai,et al.  Expanding the Detection of Traversable Area with RealSense for the Visually Impaired , 2016, Sensors.

[56]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[57]  Stephanie Ludi,et al.  Towards a situation awareness design to improve visually impaired orientation in unfamiliar buildings: Requirements elicitation study , 2014, 2014 IEEE 22nd International Requirements Engineering Conference (RE).

[58]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[59]  Adrian F. Clark,et al.  A Navigation System for the Visually Impaired: A Fusion of Vision and Depth Sensor , 2015, Applied bionics and biomechanics.

[60]  Zeyun Yu,et al.  Object Detection to Assist Visually Impaired People: A Deep Neural Network Adventure , 2018, ISVC.

[61]  Forrest N. Iandola,et al.  SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.

[62]  Laura Giarré,et al.  Enabling independent navigation for visually impaired people through a wearable vision-based feedback system , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[63]  Cang Ye,et al.  An Indoor Wayfinding System Based on Geometric Features Aided Graph SLAM for the Visually Impaired , 2017, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[64]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[65]  Kenta Imai,et al.  Detecting walkable plane areas by using RGB-D camera and accelerometer for visually impaired people , 2017, 2017 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON).

[66]  Roberta L. Klatzky,et al.  Stated Preferences for Components of a Personal Guidance System for Nonvisual Navigation , 2004 .