Depth from Defocus via Discriminative Metric Learning

In this paper, we propose a discriminative learning-based method for recovering the depth of a scene from multiple defocused images. The proposed method consists of a discriminative learning phase and a depth estimation phase. In the discriminative learning phase, we formalize depth from defocus (DFD) as a multi-class classification problem which can be solved by learning the discriminative metrics from the synthetic training set by minimizing a criterion function. To enhance the discriminative and generalization performance of the learned metrics, the criterion takes both within-class and between-class variations into account, and incorporates margin constraints. In the depth estimation phase, for each pixel, we compute the N discriminative functions and determine the depth level according to the minimum discriminant value. Experimental results on synthetic and real images show the effectiveness of our method in providing a reliable estimation of the depth of a scene.

[1]  Yoav Y. Schechner,et al.  Depth from Defocus vs. Stereo: How Different Really Are They? , 2004, International Journal of Computer Vision.

[2]  Alex Pentland,et al.  A New Sense for Depth of Field , 1985, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Hailin Jin,et al.  A Variational Approach to Shape from Defocus , 2002, ECCV.

[4]  Cordelia Schmid,et al.  TagProp: Discriminative metric learning in nearest neighbor models for image auto-annotation , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[5]  Djemel Ziou,et al.  Depth from Defocus Estimation in Spatial Domain , 2001, Comput. Vis. Image Underst..

[6]  Subhasis Chaudhuri,et al.  On defocus, diffusion and depth estimation , 2007, Pattern Recognit. Lett..

[7]  Ben Taskar,et al.  Discriminative learning of Markov random fields for segmentation of 3D scan data , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[8]  Stefano Soatto,et al.  Learning Shape from Defocus , 2002, ECCV.

[9]  Paolo Favaro,et al.  Recovering thin structures via nonlocal-means regularization with application to depth from defocus , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[10]  Stefano Soatto,et al.  A geometric approach to shape from defocus , 2005, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Subhasis Chaudhuri,et al.  A Variational Approach to Recovering Depth From Defocused Images , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Mads Nielsen,et al.  Computer Vision — ECCV 2002 , 2002, Lecture Notes in Computer Science.

[13]  Cordelia Schmid,et al.  Is that you? Metric learning approaches for face identification , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[14]  Kilian Q. Weinberger,et al.  Distance Metric Learning for Large Margin Nearest Neighbor Classification , 2005, NIPS.

[15]  Stefano Soatto,et al.  This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE IEEE TRANSACTION OF PATTERN RECO , 2022 .

[16]  Subhasis Chaudhuri,et al.  An MRF Model-Based Approach to Simultaneous Recovery of Depth and Restoration from Defocused Images , 1999, IEEE Trans. Pattern Anal. Mach. Intell..