Expansion hole filling in depth-image-based rendering using graph-based interpolation

Using texture and depth maps of a single reference viewpoint, depth-image-based rendering (DIBR) can synthesize a novel viewpoint image by translating texture pixels of the reference view to a virtual view, where synthesized pixel locations are derived from the associated depth pixel values. When the virtual viewpoint is located much closer to the 3D scene than the reference view (camera movement in the z-dimension), objects closer to the camera will increase in size in the virtual view faster than objects further away. A large increase in object size means that a patch of pixels sampled from an object surface in the reference view will be scattered to a larger spatial area, resulting in expansion holes. In this paper, we investigate the problem of identification and filling of expansion holes. We first propose a method based on depth histogram to identify missing or erroneously translated pixels as expansion holes. We then propose two techniques to fill in expansion holes with different computation complexity: i) linear interpolation, and ii) graph-based interpolation with a sparsity prior. Experimental results show that proper identification and filling of expansion holes can dramatically outperform inpainting procedure employed in VSRS 3.5 (up to 4.25dB).

[1]  Toshiaki Fujii,et al.  Free-Viewpoint TV , 2011, IEEE Signal Processing Magazine.

[2]  Guy Gilboa,et al.  Nonlocal Operators with Applications to Image Processing , 2008, Multiscale Model. Simul..

[3]  Yo-Sung Ho,et al.  Hole filling method using depth based in-painting for view synthesis in free viewpoint television and 3-D video , 2009, 2009 Picture Coding Symposium.

[4]  PROCEssIng magazInE IEEE Signal Processing Magazine , 2004 .

[5]  Dong Tian,et al.  View synthesis techniques for 3D video , 2009, Optical Engineering + Applications.

[6]  Aljoscha Smolic,et al.  Multi-View Video Plus Depth Representation and Coding , 2007, 2007 IEEE International Conference on Image Processing.

[7]  Leonard McMillan,et al.  Post-rendering 3D warping , 1997, SI3D.

[8]  U. Feige,et al.  Spectral Graph Theory , 2015 .

[9]  Ziad Al Kachouh,et al.  Fast DCT-based spatial domain interpolation of blocks in images , 2000, IEEE Trans. Image Process..

[10]  Thomas Maugey,et al.  R-D optimized auxiliary information for inpainting-based view synthesis , 2012, 2012 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON).

[11]  Oscar C. Au,et al.  Depth map compression using multi-resolution graph-based transform for depth-image-based rendering , 2012, 2012 19th IEEE International Conference on Image Processing.

[12]  Stephen P. Boyd,et al.  Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.

[13]  Béatrice Pesquet-Popescu,et al.  Depth-aided image inpainting for novel view synthesis , 2010, 2010 IEEE International Workshop on Multimedia Signal Processing.

[14]  S. Burak Gokturk,et al.  A Time-Of-Flight Depth Sensor - System Description, Issues and Solutions , 2004, 2004 Conference on Computer Vision and Pattern Recognition Workshop.

[15]  Zhaozheng Yin,et al.  Improving depth perception with motion parallax and its application in teleconferencing , 2009, 2009 IEEE International Workshop on Multimedia Signal Processing.

[16]  Stephan Reichelt,et al.  Depth cues in human visual perception and their realization in 3D displays , 2010, Defense + Commercial Sensing.

[17]  Antonio Ortega,et al.  Sparse representation of depth maps for efficient transform coding , 2010, 28th Picture Coding Symposium.

[18]  Jaejoon Lee,et al.  Edge-adaptive transforms for efficient depth map coding , 2010, 28th Picture Coding Symposium.