Extending 6-DoF VR Experience Via Multi-Sphere Images Interpolation

Three-degrees-of-freedom (3-DoF) omnidirectional imaging has been widely used in various applications ranging from street maps to 3-DoF VR live broadcasting. Although allowing for navigating viewpoints rotationally inside a virtual world, it does not provide motion parallax key for human 3D perception. Recent research mitigates this problem by introducing 3 transitional degrees of freedom (6-DoF) using multi-sphere images (MSI) which is beginning to show promises in handling occlusions and reflective objects. However, the design of MSI naturally limits the range of authentic 6-DoF experiences, as existing mechanisms for MSI rendering cannot fully utilize multi-layer information when synthesizing novel views between multiple MSIs. To tackle this problem and extend the 6-DoF range, we propose an MSI interpolation pipeline that utilizes adjacent MSIs' 3D information embedded inside their layers. In this work, we describe an MSI projection scheme along with an MSI interpolation network to predict intermediate MSIs in order to facilitate the need for extended range. We demonstrate that our system significantly improves the range of 6-DoF experience compared with other MSI-based methods. With extensive experiments, we show our algorithm outperforms state-of-the-art methods both qualitatively and quantitatively in synthesizing novel view panoramas.

[1]  Diego Gutierrez,et al.  Motion parallax for 360° RGBD video , 2019, IEEE Transactions on Visualization and Computer Graphics.

[2]  Stephen DiVerdi,et al.  Deep Multi Depth Panoramas for View Synthesis , 2020, ECCV.

[3]  Yuze He,et al.  Learning To Compose 6-DOF Omnidirectional Videos Using Multi-Sphere Images , 2021, 2021 IEEE International Conference on Image Processing (ICIP).

[4]  Gernot Riegler,et al.  Free View Synthesis , 2020, ECCV.

[5]  Alan L. Yuille,et al.  UnrealCV: Connecting Computer Vision to Unreal Engine , 2016, ECCV Workshops.

[6]  Vladlen Koltun,et al.  Photographic Image Synthesis with Cascaded Refinement Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[7]  Bertrand Chupeau,et al.  Learning Light Field Synthesis with Multi-Plane Images: Scene Encoding as a Recurrent Segmentation Task , 2020, 2020 IEEE International Conference on Image Processing (ICIP).

[8]  Paul Debevec,et al.  Welcome to light fields , 2018, SIGGRAPH 2018.

[9]  Christian Richardt,et al.  OmniPhotos , 2020 .

[10]  ALBERT PARRA POZO,et al.  An integrated 6DoF video camera and system design , 2019, ACM Trans. Graph..

[11]  Alexei A. Efros,et al.  The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[12]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[13]  Ravi Ramamoorthi,et al.  Local Light Field Fusion: Practical View Synthesis with Prescriptive Sampling Guidelines , 2019 .

[14]  Michael Goesele,et al.  The Replica Dataset: A Digital Replica of Indoor Spaces , 2019, ArXiv.

[15]  Hatem Ibrahem,et al.  Stereo Extrapolation: View synthesis with multiplane images , 2021, 2021 International Conference on Electronics, Information, and Communication (ICEIC).

[16]  Tobias Höllerer,et al.  PanoTrace: interactive 3D modeling of surround-view panoramic images in virtual reality , 2017, VRST.

[17]  Christian Richardt,et al.  Image-Based Scene Representations for Head-Motion Parallax in 360° Panoramas , 2020, Real VR.

[18]  Helio Pedrini,et al.  Adaptive Multiplane Image Generation from a Single Internet Picture , 2020, 2021 IEEE Winter Conference on Applications of Computer Vision (WACV).

[19]  James Tompkin,et al.  MatryODShka: Real-time 6DoF Video View Synthesis using Multi-Sphere Images , 2020, ECCV.

[20]  Feng Xu,et al.  Parallax360: Stereoscopic 360° Scene Representation for Head-Motion Parallax , 2018, IEEE Transactions on Visualization and Computer Graphics.

[21]  Paul Debevec,et al.  DeepView: View Synthesis With Learned Gradient Descent , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Christian Richardt,et al.  OmniPhotos: casual 360° VR photography , 2020, ACM Trans. Graph..

[23]  Pratul P. Srinivasan,et al.  NeRF , 2020, ECCV.

[24]  Petros Daras,et al.  OmniDepth: Dense Depth Estimation for Indoors Spherical Panoramas , 2018, ECCV.

[25]  Béatrice Pesquet-Popescu,et al.  Depth-aided image inpainting for novel view synthesis , 2010, 2010 IEEE International Workshop on Multimedia Signal Processing.

[26]  Zhili Chen,et al.  6-DOF VR videos with a single 360-camera , 2017, 2017 IEEE Virtual Reality (VR).

[27]  Christian Richardt,et al.  MegaParallax: Casual 360° Panoramas with Motion Parallax , 2019, IEEE Transactions on Visualization and Computer Graphics.

[28]  Graham Fyffe,et al.  Stereo Magnification: Learning View Synthesis using Multiplane Images , 2018, ArXiv.

[29]  Bernd Girod,et al.  Depth augmented stereo panorama for cinematic virtual reality with head-motion parallax , 2016, 2016 IEEE International Conference on Multimedia and Expo (ICME).

[30]  Noah Snavely,et al.  Single-View View Synthesis With Multiplane Images , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Jan Kautz,et al.  Extreme View Synthesis , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[32]  Jan Kautz,et al.  Super SloMo: High Quality Estimation of Multiple Intermediate Frames for Video Interpolation , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[33]  Christine Guillemot,et al.  Depth-based image completion for view synthesis , 2011, 2011 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON).

[34]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[35]  Shang-Hong Lai,et al.  Improved novel view synthesis from depth image with large baseline , 2008, 2008 19th International Conference on Pattern Recognition.

[36]  Petros Daras,et al.  Spherical View Synthesis for Self-Supervised 360° Depth Estimation , 2019, 2019 International Conference on 3D Vision (3DV).

[37]  Jochen Lang,et al.  Real-Time Panoramic Depth Maps from Omni-directional Stereo Images for 6 DoF Videos in Virtual Reality , 2019, 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR).

[38]  Jitendra Malik,et al.  View Synthesis by Appearance Flow , 2016, ECCV.

[39]  Jonathan T. Barron,et al.  Pushing the Boundaries of View Extrapolation With Multiplane Images , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[40]  Paul Debevec,et al.  Immersive light field video with a layered mesh representation , 2020, ACM Trans. Graph..

[41]  Paul Debevec,et al.  The making of welcome to light fields VR , 2018, SIGGRAPH Talks.

[42]  Ting-Chun Wang,et al.  Learning-based view synthesis for light field cameras , 2016, ACM Trans. Graph..

[43]  Ravi Ramamoorthi,et al.  Local light field fusion , 2019, ACM Trans. Graph..

[44]  Ravi Ramamoorthi,et al.  Learning to Synthesize a 4D RGBD Light Field from a Single Image , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[45]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[46]  Paul E. Debevec,et al.  Compositing light field video using multiplane images , 2019, SIGGRAPH Posters.