Automatic Spatially Varying Illumination Recovery of Indoor Scenes Based on a Single RGB-D Image

We propose an automatic framework to recover the illumination of indoor scenes based on a single RGB-D image. Unlike previous works, our method can recover spatially varying illumination without using any lighting capturing devices or HDR information. The recovered illumination can produce realistic rendering results. To model the geometry of the visible and invisible parts of scenes corresponding to the input RGB-D image, we assume that all objects shown in the image are located in a box with six faces and build a planar-based geometry model based on the input depth map. We then present a confidence-scoring based strategy to separate the light sources from the highlight areas. The positions of light sources both in and out of the camera's view are calculated based on the classification result and the recovered geometry model. Finally, an iterative procedure is proposed to calculate the colors of light sources and the materials in the scene. In addition, a data-driven method is used to set constraints on the light source intensities. Using the estimated light sources and geometry model, environment maps at different points in the scene are generated that can model the spatial variance of illumination. The experimental results demonstrate the validity and flexibility of our approach.

[1]  Hui-Liang Shen,et al.  Simple and efficient method for specularity removal in an image. , 2009, Applied optics.

[2]  Paul E. Debevec,et al.  Rendering synthetic objects into real scenes: bridging traditional and image-based graphics with global illumination and high dynamic range photography , 1998, SIGGRAPH '08.

[3]  Anders Ynnerman,et al.  Spatially varying image based lighting by light probe sequences , 2007, The Visual Computer.

[4]  Andrew I. Comport,et al.  3D High Dynamic Range dense visual SLAM and its application to real-time object re-lighting , 2013, 2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[5]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[6]  Jitendra Malik,et al.  Perceptual Organization and Recognition of Indoor Scenes from RGB-D Images , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[7]  Yee-Hong Yang,et al.  Multiple Illuminant Direction Detection with Application to Image Synthesis , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[8]  Anders Ynnerman,et al.  Free Form Incident Light Fields , 2008, Comput. Graph. Forum.

[9]  Kalyan Sunkavalli,et al.  Automatic Scene Inference for 3D Object Compositing , 2014, ACM Trans. Graph..

[10]  Erik Reinhard,et al.  Multiple Light Source Estimation in a Single Image , 2013, Comput. Graph. Forum.

[11]  Andrew Zisserman,et al.  Multiple View Geometry in Computer Vision (2nd ed) , 2003 .

[12]  Katsushi Ikeuchi,et al.  Illumination from Shadows , 2003, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Haibin Ling,et al.  Sparse representation of cast shadows via ℓ1-regularized least squares , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[14]  Jianxiong Xiao,et al.  SUN RGB-D: A RGB-D scene understanding benchmark suite , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Qunsheng Peng,et al.  Lighting Simulation of Augmented Outdoor Scene Based on a Legacy Photograph , 2013, Comput. Graph. Forum.

[16]  Andrew Gardner,et al.  Capturing and Rendering with Incident Light Fields , 2003, Rendering Techniques.

[17]  E. Land,et al.  Lightness and retinex theory. , 1971, Journal of the Optical Society of America.

[18]  Erik Reinhard,et al.  Image-based material editing , 2005, SIGGRAPH '05.

[19]  Andrew Gardner,et al.  Photorealistic rendering of mixed reality scenes , 2015, Comput. Graph. Forum.

[20]  David A. Forsyth,et al.  Rendering synthetic objects into legacy photographs , 2011, ACM Trans. Graph..

[21]  Masahiro Yamaguchi,et al.  High dynamic range images capturing with multi-band camera , 2007, SIGGRAPH '07.

[22]  Arnaud Darmont High Dynamic Range Imaging: Sensors and Architectures , 2013 .

[23]  P. Hanrahan,et al.  On the relationship between radiance and irradiance: determining the illumination from images of a convex Lambertian object. , 2001, Journal of the Optical Society of America. A, Optics, image science, and vision.

[24]  Tianli Yu,et al.  Sparse lumigraph relighting by illumination and reflectance estimation from multi-view images , 2006, EGSR '06.

[25]  Thomas A. Funkhouser,et al.  Semantic Scene Completion from a Single Depth Image , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  Katsushi Ikeuchi,et al.  Acquiring a Radiance Distribution to Superimpose Virtual Objects onto Real Scene , 2001, MVA.

[27]  Andrew Jones,et al.  Direct HDR capture of the sun and sky , 2006, SIGGRAPH Courses.

[28]  Hans-Peter Seidel,et al.  Relighting objects from image collections , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[29]  André Gagalowicz,et al.  Image-based rendering of diffuse, specular and glossy surfaces from a single image , 2001, SIGGRAPH.

[30]  Roberto Scopigno,et al.  EnvyDepth: An Interface for Recovering Local Natural Illumination from Environment Maps , 2013, Comput. Graph. Forum.

[31]  Andrew I. Comport,et al.  3 D High Dynamic Range Dense Visual SLAM and Its Application to Real-time Object Relighting , 2019 .

[32]  Andrew W. Fitzgibbon,et al.  KinectFusion: real-time 3D reconstruction and interaction using a moving depth camera , 2011, UIST.

[33]  Katsushi Ikeuchi,et al.  Light source position and reflectance estimation from a single view without the distant illumination assumption , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[34]  Hans-Peter Seidel,et al.  Enhancement of Bright Video Features for HDR Displays , 2008 .

[35]  Derek Hoiem,et al.  Indoor Segmentation and Support Inference from RGBD Images , 2012, ECCV.