Recent research on computer vision has made significant progress in 3d reconstruction and free-viewpoint video however most of these methods are not suited for real-time rendering. This paper presents a Video-Based Rendering method that provides online new viewpoints of the scene from a set of webcams. Our method follows a plane-sweep approach perfectly suited for GPU implementation. This paper mainly focuses on different implementations of this method for different purposes. Indeed, our basic plane-sweep technique uses 4 input cameras to create online new views. However this method can be modified to be used with up to 10 cameras or more by a camera selection process. This method can also be adapted to handle moving input cameras using a real-time camera calibration technique. Moreover, this method can easily render a depth map instead of a new view. Finally, we explain how the plane-sweep algorithm can be modified to create multiple new views simultaneously. This latter application is mainly designed for autostereoscopic displays application. Implementation and performance are detailed for all these plane-sweep methods.
[1]
Hans-Peter Seidel,et al.
Hardware-Accelerated Visual Hull Reconstruction and Rendering
,
2003,
Graphics Interface.
[2]
Matt Pharr,et al.
Gpu gems 2: programming techniques for high-performance graphics and general-purpose computation
,
2005
.
[3]
Zhengyou Zhang,et al.
A Flexible New Technique for Camera Calibration
,
2000,
IEEE Trans. Pattern Anal. Mach. Intell..
[4]
Leonard McMillan,et al.
A Real-Time Distributed Light Field Camera
,
2002,
Rendering Techniques.
[6]
Ruigang Yang,et al.
Real‐Time Consensus‐Based Scene Reconstruction Using Commodity Graphics Hardware †
,
2003,
Comput. Graph. Forum.
[7]
Robert T. Collins,et al.
A space-sweep approach to true multi-image matching
,
1996,
Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition.