Adaptive View Sampling for Efficient Synthesis of 3D View Using Calibrated Array Cameras

Recovery of three-dimensional (3D) coordinates using a set of images with texture mapping to generate a 3D mesh has been of great interest in computer graphics and 3D imaging applications. This work aims to propose an approach to adaptive view selection (AVS) that determines the optimal number of images to generate the synthesis result using the 3D mesh and textures in terms of computational complexity and image quality (peak signal-to-noise ratio (PSNR)). All 25 images were acquired by a set of cameras in a 5×5 array structure, and rectification had already been performed. To generate the mesh, depth map extraction was carried out by calculating the disparity between the matched feature points. Synthesis was performed by fully exploiting the content included in the images followed by texture mapping. Both the 2D colored images and grey-scale depth images were synthesized based on the geometric relationship between the images, and to this end, three-dimensional synthesis was performed with a smaller number of images, which was less than 25. This work determines the optimal number of images that sufficiently provides a reliable 3D extended view by generating a mesh and image textures. The optimal number of images contributes to an efficient system for 3D view generation that reduces the computational complexity while preserving the quality of the result in terms of the PSNR. To substantiate the proposed approach, experimental results are provided.

[1]  Jason Geng,et al.  Structured-light 3D surface imaging: a tutorial , 2011 .

[2]  Matthew A. Brown,et al.  Automatic Panoramic Image Stitching using Invariant Features , 2007, International Journal of Computer Vision.

[3]  Janne Heikkilä,et al.  Geometric Camera Calibration Using Circular Control Points , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[4]  Cláudio Rosito Jung,et al.  Disparity map estimation and view synthesis using temporally adaptive triangular meshes , 2017, Comput. Graph..

[5]  Gordon Wetzstein,et al.  Compressive light field photography using overcomplete dictionaries and optimized projections , 2013, ACM Trans. Graph..

[6]  In So Kweon,et al.  Depth from a Light Field Image with Learning-Based Matching Costs , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Ivo Ihrke,et al.  Principles of Light Field Imaging: Briefly revisiting 25 years of research , 2016, IEEE Signal Processing Magazine.

[8]  Deokwoo Lee,et al.  Block-Based Stereo Matching Using Image Segmentation , 2019 .

[9]  Paul Debevec,et al.  Immersive light field video with a layered mesh representation , 2020, ACM Trans. Graph..

[10]  Chongxiu Yu,et al.  Efficient DIBR method based on depth offset mapping for 3D image rendering , 2019, Applied Optics and Photonics China.

[11]  Antonios Gasteratos,et al.  Review of Stereo Vision Algorithms: From Software to Hardware , 2008 .

[12]  Li Zhang,et al.  Soft 3D reconstruction for view synthesis , 2017, ACM Trans. Graph..

[13]  Changyan Xiao,et al.  Complete grid pattern decoding method for a one-shot structured light system. , 2020, Applied optics.

[14]  Michael Weinmann,et al.  SLAMCast: Large-Scale, Real-Time 3D Reconstruction and Streaming for Immersive Multi-Client Live Telepresence , 2018, IEEE Transactions on Visualization and Computer Graphics.

[15]  Hsin-Ying Lee,et al.  Semantic View Synthesis , 2020, ECCV.

[16]  Roger Y. Tsai,et al.  A versatile camera calibration technique for high-accuracy 3D machine vision metrology using off-the-shelf TV cameras and lenses , 1987, IEEE J. Robotics Autom..

[17]  George Drettakis,et al.  Scalable inside-out image-based rendering , 2016, ACM Trans. Graph..

[18]  Menglian Zhao,et al.  A Real-Time High-Quality Complete System for Depth Image-Based Rendering on FPGA , 2019, IEEE Transactions on Circuits and Systems for Video Technology.

[19]  Xu Zhang,et al.  Robust pattern decoding in shape-coded structured light , 2017 .

[20]  Paul E. Debevec,et al.  A system for acquiring, processing, and rendering panoramic light field stills for virtual reality , 2018, ACM Trans. Graph..

[21]  Jorge L. Flores,et al.  3D-shape of objects with straight line-motion by simultaneous projection of color coded patterns , 2018 .

[22]  Steve Marschner,et al.  Matching Real Fabrics with Micro-Appearance Models , 2015, ACM Trans. Graph..

[23]  Jilai Song,et al.  Lasers structured light with phase-shifting for dense depth perception , 2019, Results in Physics.

[24]  Jonathan Richard Shewchuk,et al.  Delaunay refinement algorithms for triangular mesh generation , 2002, Comput. Geom..

[25]  Masahiro Yamaguchi Light-field and holographic three-dimensional displays [Invited]. , 2016, Journal of the Optical Society of America. A, Optics, image science, and vision.

[26]  Song Zhang,et al.  High-speed 3D shape measurement with structured light methods: A review , 2018, Optics and Lasers in Engineering.

[27]  Marc Levoy,et al.  Light field rendering , 1996, SIGGRAPH.

[28]  James Tompkin,et al.  MatryODShka: Real-time 6DoF Video View Synthesis using Multi-Sphere Images , 2020, ECCV.

[29]  Marc Levoy,et al.  Light field microscopy , 2006, ACM Trans. Graph..

[30]  Zhengyou Zhang,et al.  A Flexible New Technique for Camera Calibration , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[31]  Feng Xu,et al.  Parallax360: Stereoscopic 360° Scene Representation for Head-Motion Parallax , 2018, IEEE Transactions on Visualization and Computer Graphics.

[32]  Sam Van der Jeught,et al.  Real-time structured light profilometry: a review , 2016 .

[33]  Jean Meunier,et al.  3D Reconstruction With Time-of-Flight Depth Camera and Multiple Mirrors , 2018, IEEE Access.

[34]  Stefano Mattoccia,et al.  Reliable Fusion of ToF and Stereo Depth Driven by Confidence Measures , 2016, ECCV.

[35]  Ting-Chun Wang,et al.  Learning-based view synthesis for light field cameras , 2016, ACM Trans. Graph..

[36]  Jing Liu,et al.  Dot-coded structured light for accurate and robust 3D reconstruction. , 2020, Applied optics.