Fast interpolated cameras by combining a GPU based plane sweep with a max-flow regularisation algorithm

The work presents a method for the high speed calculation of crude depth maps. Performance and applicability are illustrated for view interpolation based on two input video streams, but the algorithm is perfectly amenable to multicamera environments. First a fast plane sweep algorithm generates the crude depth map. Speed results from hardware accelerated transformations and parallel processing available on the GPU. All computations on the graphical board are performed pixel-wise and a single pass of the sweep only processes one input resolution. A second step uses a min-cut/max-flow algorithm to ameliorate the previous result. The depth map, a noisy interpolated image and correlation measures are available on the GPU. They are reused and combined with spatial connectivity information and temporal continuity considerations in a graph formulation. Position dependent sampling densities allow the system to use multiple image resolutions. The min-cut separation of this graph yields the global minimum of the associated energy function. Limiting the search range according to the initialisation provided by the plane sweep further speeds up the process. The required hardware is only two cameras and a regular PC.

[1]  Lance Williams,et al.  View Interpolation for Image Synthesis , 1993, SIGGRAPH.

[2]  Václav Hlavác,et al.  Rendering real-world objects using view interpolation , 1995, Proceedings of IEEE International Conference on Computer Vision.

[3]  Richard Szeliski,et al.  The lumigraph , 1996, SIGGRAPH.

[4]  Steven M. Seitz,et al.  Toward image-based scene representation using view morphing , 1996, Proceedings of 13th International Conference on Pattern Recognition.

[5]  Marc Levoy,et al.  Light field rendering , 1996, SIGGRAPH.

[6]  Richard Szeliski,et al.  A layered approach to stereo reconstruction , 1998, Proceedings. 1998 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No.98CB36231).

[7]  Yizhou Yu,et al.  Efficient View-Dependent Image-Based Rendering with Projective Texture-Mapping , 1998, Rendering Techniques.

[8]  Michael Bosse,et al.  Unstructured lumigraph rendering , 2001, SIGGRAPH.

[9]  Til Aach,et al.  Illumination-Invariant Change Detection Using a Statistical Colinearity Criterion , 2001, DAGM-Symposium.

[10]  D. Scharstein,et al.  A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms , 2001, Proceedings IEEE Workshop on Stereo and Multi-Baseline Vision (SMBV 2001).

[11]  Heinrich Niemann,et al.  Dense disparity maps in real-time with an application to augmented reality , 2002, Sixth IEEE Workshop on Applications of Computer Vision, 2002. (WACV 2002). Proceedings..

[12]  Andrew Blake,et al.  Gaze manipulation for one-to-one teleconferencing , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[13]  Li Zhang,et al.  Spacetime stereo: shape recovery for dynamic scenes , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[14]  Stephen M. Smith,et al.  SUSAN—A New Approach to Low Level Image Processing , 1997, International Journal of Computer Vision.

[15]  Vladimir Kolmogorov,et al.  An experimental comparison of min-cut/max- flow algorithms for energy minimization in vision , 2001, IEEE Transactions on Pattern Analysis and Machine Intelligence.