Calibrationless Sensor Fusion Using Linear Optimization for Depth Matching

Recently the observation of surveillanced areas scanned by multi-camera systems is getting more and more popular. The newly developed sensors give new opportunities for exploiting novel features. Using the information gained from a conventional camera we have data about the colours, the shape of objects and the micro-structures; and we have additional information while using thermal camera in the darkness. A camera with depth sensor can find the motion and the position of an object in space even in the case when conventional cameras are unusable. How can we register the corresponding elements on different pictures? There are numerous approaches to the solution of this problem. One of the most used solutions is that the registration is based on the motion. In this method it is not necessary to look for the main features on the pictures to register the related objects, since the features would be different because of the different properties of the cameras. It is easier and faster if the registration is based on the motion. But other problems will arise in this case: shadows or shiny specular surfaces cause problems at the motion. This paper is about how we can register the corresponding elements in a multi-camera system, and how we can find a homography between the image planes in real time, so that we can register a moving object in the images of different cameras based on the depth information.

[1]  Marc Van Droogenbroeck,et al.  Background subtraction: Experiments and improvements for ViBe , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[2]  A. Ruszczynski,et al.  Nonlinear Optimization , 2006 .

[3]  Mubarak Shah,et al.  Tracking Multiple Occluding People by Localizing on Multiple Scene Planes , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  László Havasi,et al.  Stochastic View Registration of Overlapping Cameras Based on Arbitrary Motion , 2007, IEEE Transactions on Image Processing.

[5]  Carsten Rother Linear multiview reconstruction of points, lines, planes and cameras using a reference plane , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[6]  Paul A. Beardsley,et al.  Sensor Fusion for Depth Estimation, including TOF and Thermal Sensors , 2012, 2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission.

[7]  Richard I. Hartley,et al.  Theory and Practice of Projective Rectification , 1999, International Journal of Computer Vision.

[8]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[9]  Andrew Blake,et al.  Efficient Human Pose Estimation from Single Depth Images , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[11]  Ian D. Reid,et al.  Single View Metrology , 2000, International Journal of Computer Vision.

[12]  Marc Van Droogenbroeck,et al.  ViBe: A Universal Background Subtraction Algorithm for Video Sequences , 2011, IEEE Transactions on Image Processing.

[13]  Marc Van Droogenbroeck,et al.  ViBE: A powerful random technique to estimate the background in video sequences , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.