Depth from Defocus via Active Multispectral Quasi-random Point Projections using Deep Learning

A novel approach for inferring depth measurements via multispectral active depth from defocus and deep learning has been designed, implemented, and successfully tested. The scene is actively illuminated with a multispectral quasi-random point pattern, and a conventional RGB camera is used to acquire images of the projected pattern. The projection points in the captured image of the projected pattern are analyzed using an ensemble of deep neural networks to estimate the depth at each projection point. A final depth map is then reconstructed algorithmically based on the point depth estimates. Experiments using different test scenes with different structural characteristics show that the proposed approach can produced improved depth maps compared to prior deep learning approaches using monospectral projection patterns.

[1]  A. Ma,et al.  Enhanced depth from defocus via active quasi-random colored point projections , 2017 .

[2]  H. Niederreiter Point sets and sequences with small discrepancy , 1987 .

[3]  Alex Pentland,et al.  Simple range cameras based on focal error , 1994 .

[4]  Paul F. Whelan,et al.  Computational approach for depth from defocus , 2005, J. Electronic Imaging.

[5]  Steven A. Shafer,et al.  Depth from focusing and defocusing , 1993, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[6]  Liu Jianzhuang,et al.  Automatic thresholding of gray-level pictures using two-dimension Otsu method , 1991, China., 1991 International Conference on Circuits and Systems.

[7]  Joaquim Salvi,et al.  Pattern codification strategies in structured light systems , 2004, Pattern Recognit..

[8]  David A. Clausi,et al.  Depth from Defocus via Active Quasi-random Point Projections: A Deep Learning Approach , 2017, ICIAR.

[9]  M. Turk,et al.  A simple, real-time range camera , 1989, Proceedings CVPR '89: IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[10]  A. Ma,et al.  Depth from Defocus via Active Quasi-random Point Projections , 2016 .

[11]  Shree K. Nayar,et al.  Rational Filters for Passive Depth from Defocus , 1998, International Journal of Computer Vision.

[12]  Joachim Hertzberg,et al.  An autonomous mobile robot with a 3D laser range finder for 3D exploration and digitalization of indoor environments , 2003, Robotics Auton. Syst..

[13]  Robert Bridson,et al.  Fast Poisson disk sampling in arbitrary dimensions , 2007, SIGGRAPH '07.

[14]  P. Belhumeur,et al.  Active refocusing of images and videos , 2007, ACM Trans. Graph..