View Synthesis Based on Background Update with Gaussian Mixture Model

View Synthesis is a key technique for 3-D video and free view video generation. In the traditional 3-D video and free view video, there have to be many real cameras to capture the scene at the large cost. With the help of the view synthesis technique, a limited number of cameras can achieve the goal of multi-view generation. However, some holes will appear in the synthesized views due to the 3-D warping process in the view synthesis system. These holes seriously affect the quality of the synthesized images, especially for the disocclusions which is caused from the occluded regions in the original view may become visible in the virtual view. In this paper, we focus on the disocclustion filling after 3-D warping in view synthesis system. An approach is proposed to fill the discocclusion by using the real background information covered in the original view, which is based on the observation that the covered information in the current frame may be visible in the next frames of the same view. In this approach, the stable texture and depth background reference frames are generated for the left and right view, respectively, which are based on the Gaussian Mixture Model (GMM). Then, in the view synthesis system, a stable background reference frame is merged by the left and right warped images with the corresponding texture and depth background reference frames. Finally, the merged frame is used to fill the disocclusion regions of each merged frame as the background reference frame. The experimental results show that the proposed scheme can achieve better objective quality, especially for the scene with moving objects.

[1]  Masayuki Tanimoto,et al.  Multiview Imaging and 3DTV , 2007, IEEE Signal Processing Magazine.

[2]  Dong Tian,et al.  Boundary Artifact Reduction in View Synthesis of 3D Video: From Perspective of Texture-Depth Alignment , 2011, IEEE Transactions on Broadcasting.

[3]  Rainer Stiefelhagen,et al.  Evaluating Multiple Object Tracking Performance: The CLEAR MOT Metrics , 2008, EURASIP J. Image Video Process..

[4]  Liang Zhang,et al.  Stereoscopic image generation based on depth images for 3D TV , 2005, IEEE Transactions on Broadcasting.

[5]  Thomas Wiegand,et al.  Temporally consistent handling of disocclusions with texture synthesis for depth-image-based rendering , 2010, 2010 IEEE International Conference on Image Processing.

[6]  W. Eric L. Grimson,et al.  Adaptive background mixture models for real-time tracking , 1999, Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149).

[7]  Dar-Shyang Lee,et al.  Effective Gaussian mixture learning for video background subtraction , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  Aljoscha Smolic,et al.  View Synthesis for Advanced 3D Video Systems , 2008, EURASIP J. Image Video Process..

[9]  Demin Wang,et al.  Stereoscopic image generation based on depth images , 2004, 2004 International Conference on Image Processing, 2004. ICIP '04..

[10]  Manoranjan Paul,et al.  Improved Gaussian mixtures for robust object detection by adaptive multi-background generation , 2008, 2008 19th International Conference on Pattern Recognition.

[11]  Liang-Gee Chen,et al.  Efficient Depth Image Based Rendering with Edge Dependent Depth Filter and Interpolation , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[12]  Pei-Jun Lee,et al.  Nongeometric Distortion Smoothing Approach for Depth Map Preprocessing , 2011, IEEE Transactions on Multimedia.

[13]  Ismo Rakkolainen,et al.  A Survey of 3DTV Displays: Techniques and Technologies , 2007, IEEE Transactions on Circuits and Systems for Video Technology.