Line image signature for scene understanding with a wearable vision system

Wearable computer vision systems provide plenty of opportunities to develop human assistive devices. This work contributes on visual scene understanding techniques using a helmet-mounted omnidirectional vision system. The goal is to extract semantic information of the environment, such as the type of environment being traversed or the basic 3D layout of the place, to build assistive navigation systems. We propose a novel line-based image global descriptor that encloses the structure of the scene observed. This descriptor is designed with omnidirectional imagery in mind, where observed lines are longer than in conventional images. Our experiments show that the proposed descriptor can be used for indoor scene recognition comparing its results to state-of-the-art global descriptors. Besides, we demonstrate additional advantages of particular interest for wearable vision systems: higher robustness to rotation, compactness, and easier integration with other scene understanding steps.

[1]  Antonio Torralba,et al.  Modeling the Shape of the Scene: A Holistic Representation of the Spatial Envelope , 2001, International Journal of Computer Vision.

[2]  Ana Cristina Murillo,et al.  From lines to epipoles through planes in two views , 2006, Pattern Recognit..

[3]  Jesus Bermudez-Cameo,et al.  Hypercatadioptric Line Images for 3D Orientation and Image Rectification , 1970 .

[4]  Bill Triggs,et al.  Histograms of oriented gradients for human detection , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[5]  Alexei A. Efros,et al.  Segmenting Scenes by Matching Image Composites , 2009, NIPS.

[6]  Lu Wang,et al.  Wide-baseline image matching using Line Signatures , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[7]  Jitendra Malik,et al.  Shape matching and object recognition using shape contexts , 2010, 2010 3rd International Conference on Computer Science and Information Technology.

[8]  Pascal Vasseur,et al.  Rotation estimation and vanishing point extraction by omnidirectional vision in urban environment , 2012, Int. J. Robotics Res..

[9]  Marc Alexa,et al.  Sketch-based 3D shape retrieval , 2010, SIGGRAPH '10.

[10]  Kostas Daniilidis,et al.  A Unifying Theory for Central Panoramic Systems and Practical Applications , 2000, ECCV.

[11]  Josechu J. Guerrero,et al.  Omnidirectional Vision for Indoor Spatial Layout Recovery , 2013, Frontiers of Intelligent Autonomous Systems.

[12]  Wei Zhang,et al.  Video Compass , 2002, ECCV.

[13]  Marc Pollefeys,et al.  Efficient Structured Prediction with Latent Variables for General Graphical Models , 2012, ICML.

[14]  J. Kosecka,et al.  Gist vocabularies in omnidirectional images for appearance based mapping and localization , 2010 .

[15]  Alejandro Rituerto,et al.  Semantic labeling for indoor topological mapping using a wearable catadioptric system , 2014, Robotics Auton. Syst..

[16]  Derek Hoiem,et al.  Recovering the spatial layout of cluttered rooms , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[17]  Luc Van Gool,et al.  Wide-baseline stereo matching with line segments , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[18]  Zhanyi Hu,et al.  MSLD: A robust descriptor for line matching , 2009, Pattern Recognit..

[19]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[20]  Frédéric Jurie,et al.  Groups of Adjacent Contour Segments for Object Detection , 2008, IEEE Trans. Pattern Anal. Mach. Intell..