Improving Indoor Mobility of the Visually Impaired with Depth-Based Spatial Sound

We present a novel system to help visually impaired people to move efficiently and safely in indoor environments by mapping input from a depth camera to spatially localized auditory cues. We propose a set of context-specific cues which are suitable for use in systems that provide minimal audio feedback and hence reduce masking of natural sounds compared to the audio provided by general-purpose sense substitution devices. Using simple but effective heuristics for detecting the floor and the side walls, we propose auditory cues that encode information about the distances to walls, obstacles, the orientation of the corridor or room, and openings into corridors or rooms. But the key to our system is the use of a spatial sound engine that localizes the generated sounds in 3D. We evaluate our system, comparing with MeloSee. Our preliminary pilot study with ten blindfolded participants suggests that our system was more helpful for spotting smaller obstacles on the floor, though neither system had a significant edge in terms of walking speed or safety.

[1]  João Barroso,et al.  Blind Navigation Support System based on Microsoft Kinect , 2012, DSAI.

[2]  Guido Bologna,et al.  Transforming 3D Coloured Pixels into Musical Instrument Notes for Vision Substitution Applications , 2007, EURASIP J. Image Video Process..

[3]  John C. Platt,et al.  HRTF magnitude synthesis via sparse representation of anthropometric features , 2014, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[4]  Ivan Tashev,et al.  HRTF magnitude modeling using a non-regularized least-squares fit of spherical harmonics coefficients on incomplete data , 2012, Proceedings of The 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference.

[5]  Gérard G. Medioni,et al.  Wearable RGBD Indoor Navigation System for the Blind , 2014, ECCV Workshops.

[6]  Susan M. Downes,et al.  A Depth-Based Head-Mounted Visual Display to Aid Navigation in Partially Sighted Individuals , 2013, PloS one.

[7]  Eduardo Bayro-Corrochano,et al.  Geometric fuzzy techniques for guidance of visually impaired people , 2013 .

[8]  Denis Pellerin,et al.  Depth — Melody substitution , 2012, 2012 Proceedings of the 20th European Signal Processing Conference (EUSIPCO).

[9]  M. Paré,et al.  Early-blind human subjects localize sound sources better than sighted subjects , 1998, Nature.

[10]  Denis Pellerin,et al.  Navigating from a Depth Image Converted into Sound , 2015, Applied bionics and biomechanics.

[11]  Ivan Tashev,et al.  HRTF phase synthesis via sparse representation of anthropometric features , 2014, 2014 Information Theory and Applications Workshop (ITA).

[12]  Shyamanta M. Hazarika,et al.  IntelliNavi: Navigation for Blind Based on Kinect and Machine Learning , 2014, MIWAI.

[13]  James R. Marston,et al.  Cognitive load of navigating without vision when guided by virtual sound versus spatial language. , 2006, Journal of experimental psychology. Applied.

[14]  Peter B. L. Meijer,et al.  An experimental system for auditory image representations , 1992, IEEE Transactions on Biomedical Engineering.

[15]  Myung Jin Chung,et al.  Real-time plane detection based on depth map from Kinect , 2013, IEEE ISR 2013.

[16]  José L. González-Mora,et al.  Development of a New Space Perception System for Blind People, Based on the Creation of a Virtual Acoustic Space , 1999, IWANN.

[17]  Per Ola Kristensson,et al.  Supporting blind navigation using depth sensing and sonification , 2013, UbiComp.