An Approach to Distance Estimation with Stereo Vision Using Address-Event-Representation

Image processing in digital computer systems usually considers the visual information as a sequence of frames. These frames are from cameras that capture reality for a short period of time. They are renewed and transmitted at a rate of 25-30 fps (typical real-time scenario). Digital video processing has to process each frame in order to obtain a result or detect a feature. In stereo vision, existing algorithms used for distance estimation use frames from two digital cameras and process them pixel by pixel to obtain similarities and differences from both frames; after that, depending on the scene and the features extracted, an estimate of the distance of the different objects of the scene is calculated. Spike-based processing is a relatively new approach that implements the processing by manipulating spikes one by one at the time they are transmitted, like a human brain. The mammal nervous system is able to solve much more complex problems, such as visual recognition by manipulating neuron spikes. The spike-based philosophy for visual information processing based on the neuro-inspired Address-Event-Representation (AER) is achieving nowadays very high performances. In this work we propose a two-DVS-retina system, composed of other elements in a chain, which allow us to obtain a distance estimation of the moving objects in a close environment. We will analyze each element of this chain and propose a Multi Hold&Fire algorithm that obtains the differences between both retinas.

[1]  T R Crimmins,et al.  Geometric filter for speckle reduction. , 1985, Applied optics.

[2]  Angel Jiménez-Fernandez,et al.  Building blocks for spikes signals processing , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).

[3]  Wayne Luk,et al.  Have GPUs made FPGAs redundant in the field of video processing? , 2005, Proceedings. 2005 IEEE International Conference on Field-Programmable Technology, 2005..

[4]  M. Domínguez-Morales,et al.  Image matching algorithms in stereo vision using address-event-representation: A theoretical study and evaluation of the different algorithms , 2011, Proceedings of the International Conference on Signal Processing and Multimedia Applications.

[5]  Ryad Benosman,et al.  Panoramic stereo vision sensor , 1998, Proceedings. Fourteenth International Conference on Pattern Recognition (Cat. No.98EX170).

[6]  Massimo A. Sivilotti,et al.  Wiring considerations in analog VLSI systems, with application to field-programmable networks , 1992 .

[7]  Tobi Delbrück,et al.  A 128$\times$ 128 120 dB 15 $\mu$s Latency Asynchronous Temporal Contrast Vision Sensor , 2008, IEEE Journal of Solid-State Circuits.

[8]  Tobi Delbrück,et al.  A 5 Meps $100 USB2.0 Address-Event Monitor-Sequencer Interface , 2007, 2007 IEEE International Symposium on Circuits and Systems.

[9]  Ryad Benosman,et al.  A multi-cameras 3D volumetric method for outdoor scenes: a road traffic monitoring application , 2004, ICPR 2004.

[10]  Martin A. Fischler,et al.  Computational Stereo , 1982, CSUR.

[11]  R. Benosman,et al.  Real time omni-directional stereovision and planes detection , 1996, Proceedings of 8th Mediterranean Electrotechnical Conference on Industrial Applications in Power Systems, Computer Science and Telecommunications (MELECON 96).

[12]  Ryad Benosman,et al.  Panoramic stereovision sensor , 1998 .

[13]  G. Shepherd The Synaptic Organization of the Brain , 1979 .

[14]  Jong-Sen Lee,et al.  A simple speckle smoothing algorithm for synthetic aperture radar images , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[15]  Angel Jiménez-Fernandez,et al.  On the AER convolution processors for FPGA , 2010, Proceedings of 2010 IEEE International Symposium on Circuits and Systems.