On-Line Novel View Synthesis Capable of Handling Multiple Moving Objects

This paper presents a new interactive teleconferencing system. It adds a ‘virtual' camera to the scene which can move freely in between multiple real cameras. The viewpoint can automatically be selected using basic cinematographic rules, based on the position and the actions of the instructor. This produces a clearer and more engaging view for the remote audience, without the need for a human editor. For the creation of the novel views generated by such a ‘virtual' camera, segmentation and depth calculations are required. The system is semi-automatic, in that the user is asked to indicate a few corresponding points or edges for generating an initial rough background model. Next to the static background and moving foreground also multiple independently moving objects are catered for. The initial foreground contour is tracked over time, using a new active contour. If a second object appears, the contour prediction allows to recognize this situation and to take appropriate measures. The 3D models are continuously validated based on a Birchfield dissimilarity measure. The foreground model is updated every frame, the background is refined if necessary. The current implementation can reach approx 4 fps on a single desktop.

[1]  Luc Van Gool,et al.  Hierarchical coarse to fine depth estimation for realistic view interpolation , 2005, Fifth International Conference on 3-D Digital Imaging and Modeling (3DIM'05).

[2]  Carlo Tomasi,et al.  A Pixel Dissimilarity Measure That Is Insensitive to Image Sampling , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  Hirokazu Kato,et al.  Real World Teleconferencing , 2002, IEEE Computer Graphics and Applications.

[4]  Ruigang Yang,et al.  Real-time view synthesis using commodity graphics hardware , 2002, SIGGRAPH '02.

[5]  Andrew Blake,et al.  Gaze manipulation for one-to-one teleconferencing , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.

[6]  Reinhard Koch,et al.  Self-Calibration and Metric Reconstruction Inspite of Varying and Unknown Intrinsic Camera Parameters , 1999, International Journal of Computer Vision.

[7]  Bill Triggs,et al.  Autocalibration and the absolute quadric , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[8]  L. Matthies,et al.  Enhanced real-time stereo using bilateral filtering , 2004 .

[9]  Dimitri Konstantas,et al.  Design and Implementation of a Teleteaching Environment , 1999 .

[10]  Mark Billinghurst,et al.  A wearable spatial conferencing space , 1998, Digest of Papers. Second International Symposium on Wearable Computers (Cat. No.98EX215).

[11]  Antonio Criminisi,et al.  The SPS algorithm: patching figural continuity and transparency by Split-Patch Search , 2004, CVPR 2004.

[12]  Ruigang Yang,et al.  Real-time consensus-based scene reconstruction using commodity graphics hardware , 2002, 10th Pacific Conference on Computer Graphics and Applications, 2002. Proceedings..

[13]  Hirokazu Kato,et al.  Real world teleconferencing , 1999, CHI Extended Abstracts.

[14]  Jonathan M. Garibaldi,et al.  Real-Time Correlation-Based Stereo Vision with Reduced Border Errors , 2002, International Journal of Computer Vision.