Single-photon 3D imaging with deep sensor fusion

Sensors which capture 3D scene information provide useful data for tasks in vehicle navigation, gesture recognition, human pose estimation, and geometric reconstruction. Active illumination time-of-flight sensors in particular have become widely used to estimate a 3D representation of a scene. However, the maximum range, density of acquired spatial samples, and overall acquisition time of these sensors is fundamentally limited by the minimum signal required to estimate depth reliably. In this paper, we propose a data-driven method for photon-efficient 3D imaging which leverages sensor fusion and computational reconstruction to rapidly and robustly estimate a dense depth map from low photon counts. Our sensor fusion approach uses measurements of single photon arrival times from a low-resolution single-photon detector array and an intensity image from a conventional high-resolution camera. Using a multi-scale deep convolutional network, we jointly process the raw measurements from both sensors and output a high-resolution depth map. To demonstrate the efficacy of our approach, we implement a hardware prototype and show results using captured data. At low signal-to-background levels, our depth reconstruction algorithm with sensor fusion outperforms other methods for depth estimation from noisy measurements of photon arrival times.

[1]  Vivek K Goyal,et al.  First-Photon Imaging , 2014, Science.

[2]  Narendra Ahuja,et al.  Deep Joint Image Filtering , 2016, ECCV.

[3]  Michael S. Brown,et al.  High quality depth map upsampling for 3D-TOF cameras , 2011, 2011 International Conference on Computer Vision.

[4]  Xiaoou Tang,et al.  Depth Map Super-Resolution by Deep Multi-Scale Guidance , 2016, ECCV.

[5]  Sebastian Thrun,et al.  An Application of Markov Random Fields to Range Sensing , 2005, NIPS.

[6]  Xiangyu Zhang,et al.  Large Kernel Matters — Improve Semantic Segmentation by Global Convolutional Network , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Gordon Wetzstein,et al.  Reconstructing Transient Images from Single-Photon Sensors , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Christopher Joseph Pal,et al.  Learning Conditional Random Fields for Stereo , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[9]  Michael Werman,et al.  Fusing Time-of-Flight Depth and Color for Real-Time Segmentation and Tracking , 2009, Dyn3D.

[10]  Conrad C. Huang,et al.  UCSF Chimera—A visualization system for exploratory research and analysis , 2004, J. Comput. Chem..

[11]  Aongus McCarthy,et al.  Long-range depth profiling of camouflaged targets using single-photon detection , 2017 .

[12]  Gordon Wetzstein,et al.  Towards transient imaging at interactive rates with single-photon detectors , 2018, 2018 IEEE International Conference on Computational Photography (ICCP).

[13]  Ruigang Yang,et al.  Spatial-Depth Super Resolution for Range Images , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[14]  Matthew O'Toole,et al.  Homogeneous codes for energy-efficient illumination and imaging , 2015, ACM Trans. Graph..

[15]  Vladlen Koltun,et al.  A Simple Model for Intrinsic Image Decomposition with Depth Cues , 2013, 2013 IEEE International Conference on Computer Vision.

[16]  R. Mcintyre,et al.  Photon counting techniques with silicon avalanche photodiodes. , 1993, Applied optics.

[17]  Steve McLaughlin,et al.  A Bayesian Approach to Denoising of Single-Photon Binary Images , 2016, IEEE Transactions on Computational Imaging.

[18]  William Whittaker,et al.  Epipolar time-of-flight imaging , 2017, ACM Trans. Graph..

[19]  D. Renker Geiger-mode avalanche photodiodes, history, properties and problems , 2006 .

[20]  Gordon Wetzstein,et al.  Deep End-to-End Time-of-Flight Imaging , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[21]  Vivek K. Goyal,et al.  A Few Photons Among Many: Unmixing Signal and Noise for Photon-Efficient Active Imaging , 2016, IEEE Transactions on Computational Imaging.

[22]  Michael F. Cohen,et al.  Digital photography with flash and no-flash image pairs , 2004, ACM Trans. Graph..

[23]  Horst Bischof,et al.  Image Guided Depth Upsampling Using Anisotropic Total Generalized Variation , 2013, 2013 IEEE International Conference on Computer Vision.

[24]  Vivek K Goyal,et al.  Photon-efficient imaging with a single-photon camera , 2016, Nature Communications.

[25]  HoraudRadu,et al.  An overview of depth cameras and range scanners based on time-of-flight technologies , 2016 .

[26]  Derek Hoiem,et al.  Indoor Segmentation and Support Inference from RGBD Images , 2012, ECCV.

[27]  Michael L. Lightstone,et al.  A new efficient approach for the removal of impulse noise from highly corrupted images , 1996, IEEE Trans. Image Process..

[28]  Dani Lischinski,et al.  Joint bilateral upsampling , 2007, ACM Trans. Graph..

[29]  Reinhard Koch,et al.  Time-of-Flight Sensors in Computer Graphics , 2009, Eurographics.

[30]  G. Buller,et al.  Kilometer-range depth imaging at 1,550 nm wavelength using an InGaAs/InP single-photon avalanche diode detector. , 2013, Optics express.

[31]  Dieter Fox,et al.  RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments , 2012, Int. J. Robotics Res..

[32]  Abderrahim Halimi,et al.  Single-photon three-dimensional imaging at up to 10 kilometers range. , 2017, Optics express.

[33]  Vivek K. Goyal,et al.  Photon-efficient computational imaging with a single-photon camera , 2016 .

[34]  Risto A. Myllylae,et al.  Comparison of continuous-wave and pulsed time-of-flight laser range-finding techniques , 1992, Other Conferences.

[35]  Min H. Kim,et al.  DeepToF: off-the-shelf real-time correction of multipath interference in time-of-flight imaging , 2017, ACM Trans. Graph..

[36]  Vivek K. Goyal,et al.  Photon-Efficient Computational 3-D and Reflectivity Imaging With Single-Photon Detectors , 2014, IEEE Transactions on Computational Imaging.

[37]  Sebastian Thrun,et al.  A Noise‐aware Filter for Real‐time Depth Upsampling , 2008 .

[38]  Ian D. Reid,et al.  RefineNet: Multi-path Refinement Networks for High-Resolution Semantic Segmentation , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[39]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[40]  Edoardo Charbon,et al.  LinoSPAD: a time-resolved 256×1 CMOS SPAD line sensor system featuring 64 FPGA-based TDC channels running at up to 8.5 giga-events per second , 2016, Photonics Europe.

[41]  Michael Wahl,et al.  Time-Correlated Single Photon Counting , 2009 .