Mobile Panoramic Vision for Assisting the Blind via Indexing and Localization

In this paper, we propose a first-person localization and navigation system for helping blind and visually-impaired people navigate in indoor environments. The system consists of a mobile vision front end with a portable panoramic lens mounted on a smart phone, and a remote GPU-enabled server. Compact and effective omnidirectional video features are extracted and represented in the smart phone front end, and then transmitted to the server, where the features of an input image or a short video clip are used to search a database of an indoor environment via image-based indexing to find both the location and the orientation of the current view. To deal with the high computational cost in searching a large database for a realistic navigation application, data parallelism and task parallelism properties are identified in database indexing, and computation is accelerated by using multi-core CPUs and GPUs. Experiments on synthetic data and real data are carried out to demonstrate the capacity of the proposed system, with respect to real-time response and robustness.

[1]  Anil A. Bharath,et al.  Mobile Visual Assistive Apps: Benchmarks of Vision Algorithm Performance , 2013, ICIAP Workshops.

[2]  John Nicholson,et al.  RFID in robot-assisted indoor navigation for the visually impaired , 2004, 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566).

[3]  Mario Innocenti,et al.  An Entropy-like approach to vision based autonomous navigation , 2011, 2011 IEEE International Conference on Robotics and Automation.

[4]  Gordon E. Legge,et al.  Indoor Navigation by People with Visual Impairment Using a Digital Sign System , 2013, PloS one.

[5]  Roberto Manduchi,et al.  The last meter: blind visual guidance to a target , 2014, CHI.

[6]  Serge J. Belongie,et al.  Recognizing locations with Google Glass: A case study , 2014, IEEE Winter Conference on Applications of Computer Vision.

[7]  Annalisa Milella,et al.  RFID tag localization by using adaptive neuro-fuzzy inference for mobile robot applications , 2012, Ind. Robot.

[8]  Zhigang Zhu,et al.  Visual noun navigation framework for the blind , 2013 .

[9]  Zhigang Zhu,et al.  Visual Nouns for Indoor/Outdoor Navigation , 2012, ICCHP.

[10]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Paul Newman,et al.  Appearance-only SLAM at large scale with FAB-MAP 2.0 , 2011, Int. J. Robotics Res..

[12]  Mohamed Aly,et al.  Street view goes indoors: Automatic pose estimation from uncalibrated unordered spherical panoramas , 2012, 2012 IEEE Workshop on the Applications of Computer Vision (WACV).

[13]  Xueyin Lin,et al.  Fast road classification and orientation estimation using omni-view images and neural networks , 1998, IEEE Trans. Image Process..

[14]  Feng Bing-shu Study on fisheye image correction based on cylinder model , 2008 .

[15]  Roland Siegwart,et al.  A Flexible Technique for Accurate Omnidirectional Camera Calibration and Structure from Motion , 2006, Fourth IEEE International Conference on Computer Vision Systems (ICVS'06).

[16]  Jana Kosecka,et al.  Localization in Urban Environments Using a Panoramic Gist Descriptor , 2013, IEEE Transactions on Robotics.