Computational Light Field Generation Using Deep Nonparametric Bayesian Learning

In this paper, we present a deep nonparametric Bayesian method to synthesize a light field from a single image. Conventionally, light-field capture requires special optical architecture, and the gain in angular resolution often comes at the expense of a reduction in spatial resolution. Techniques for computationally generating the light field from a single image can be expanded further to a variety of applications, ranging from microscopy and materials analysis to vision-based robotic control and autonomous vehicles. We treat the light field as multiple sub-aperture views, and to compute the novel viewpoints, our model contains three major components. First, a convolutional neural network is used for predicting the depth probability map from the image. Second, a multi-scale feature dictionary is constructed within a multi-layer dictionary learning network. Third, the novel views are synthesized taking into account both the probabilistic depth map and the multi-scale feature dictionary. The experiments show that our method outperforms several state-of-the-art novel view synthesis methods in delivering good image resolution.

[1]  Edmund Y. Lam,et al.  Learning-based nonparametric autofocusing for digital holography , 2018 .

[2]  Dani Lischinski,et al.  Colorization using optimization , 2004, ACM Trans. Graph..

[3]  Hayden Kwok-Hay So,et al.  Computationally Efficient Hyperspectral Data Learning Based on the Doubly Stochastic Dirichlet Process , 2017, IEEE Transactions on Geoscience and Remote Sensing.

[4]  Gordon Wetzstein,et al.  A Wide-Field-of-View Monocentric Light Field Camera , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Xiaoou Tang,et al.  Single Image Haze Removal Using Dark Channel Prior , 2011 .

[6]  Edmund Y. Lam,et al.  Data-driven light field depth estimation using deep Convolutional Neural Networks , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[7]  Edmund Y Lam,et al.  Computational photography with plenoptic camera and light field capture: tutorial. , 2015, Journal of the Optical Society of America. A, Optics, image science, and vision.

[8]  Andrew Zisserman,et al.  Deep Features for Text Spotting , 2014, ECCV.

[9]  Alexei A. Efros,et al.  A 4D Light-Field Dataset and CNN Architectures for Material Recognition , 2016, ECCV.

[10]  M. Levoy,et al.  Light field microscopy , 2006, SIGGRAPH 2006.

[11]  Gordon Wetzstein,et al.  Compressive light field photography using overcomplete dictionaries and optimized projections , 2013, ACM Trans. Graph..

[12]  P. Hanrahan,et al.  Light Field Photography with a Hand-held Plenoptic Camera , 2005 .

[13]  Edmund Y. Lam,et al.  Sparse Hierarchical Nonparametric Bayesian learning for light field representation and denoising , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[14]  Peter I. Corke,et al.  Image-Based Visual Servoing With Light Field Cameras , 2017, IEEE Robotics and Automation Letters.

[15]  Edmund Y Lam,et al.  High-resolution lightfield photography using two masks. , 2012, Optics express.

[16]  Sven Wanner,et al.  Variational Light Field Analysis for Disparity Estimation and Super-Resolution , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  P. Hanrahan,et al.  Digital light field photography , 2006 .

[18]  Chong Wang,et al.  Decoupling Sparsity and Smoothness in the Discrete Hierarchical Dirichlet Process , 2009, NIPS.

[19]  Deqing Sun,et al.  Blind Image Deblurring Using Dark Channel Prior , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Edmund Y. Lam,et al.  A high-resolution lightfield camera with dual-mask design , 2012, Other Conferences.

[21]  Ashok Veeraraghavan,et al.  Light field denoising, light field superresolution and stereo camera based refocussing using a GMM light field patch prior , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[22]  Edmund Y. Lam,et al.  Unsupervised Tracking With the Doubly Stochastic Dirichlet Process Mixture Model , 2016, IEEE Transactions on Intelligent Transportation Systems.

[23]  Alan L. Yuille,et al.  Towards unified depth and semantic prediction from a single image , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Haibin Ling,et al.  Saliency Detection on Light Field , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  M. Escobar,et al.  Markov Chain Sampling Methods for Dirichlet Process Mixture Models , 2000 .

[26]  Guosheng Lin,et al.  Deep convolutional neural fields for depth estimation from a single image , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  E. Lam,et al.  Light Field Superresolution Reconstruction in Computational Photography , 2011 .

[28]  David B. Dunson,et al.  Nonparametric Bayesian Dictionary Learning for Analysis of Noisy and Incomplete Images , 2012, IEEE Transactions on Image Processing.

[29]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[30]  Chao Sun,et al.  Synthesis of light-field raw data from RGB-D images , 2015, 2015 IEEE International Conference on Image Processing (ICIP).

[31]  Nianyi Li,et al.  A weighted sparse coding framework for saliency detection , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  Zhou Wang,et al.  Multiscale structural similarity for image quality assessment , 2003, The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003.

[33]  Edmund Y. Lam,et al.  Analysis of the noise in back-projection light field acquisition and its optimization , 2017, Applied optics.

[34]  Hao Gao,et al.  Depth-Assisted Full Resolution Network for Single Image-Based View Synthesis , 2017, IEEE Computer Graphics and Applications.

[35]  Derek Hoiem,et al.  Indoor Segmentation and Support Inference from RGBD Images , 2012, ECCV.

[36]  Ali Farhadi,et al.  Deep3D: Fully Automatic 2D-to-3D Video Conversion with Deep Convolutional Neural Networks , 2016, ECCV.

[37]  Marc Levoy,et al.  High performance imaging using large camera arrays , 2005, SIGGRAPH 2005.

[38]  Byoungho Lee,et al.  3D Imaging Based on Depth Measurement Technologies , 2018, Sensors.

[39]  Edmund Y. Lam,et al.  End-to-end deep learning framework for digital holographic reconstruction , 2019, Advanced Photonics.

[40]  Marc Levoy,et al.  Light field rendering , 1996, SIGGRAPH.

[41]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[42]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[43]  Thomas Brox,et al.  FlowNet: Learning Optical Flow with Convolutional Networks , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[44]  David B. Dunson,et al.  Deep Learning with Hierarchical Convolutional Factor Analysis , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[45]  Edmund Y Lam,et al.  High-resolution Fourier hologram synthesis from photographic images through computing the light field. , 2016, Applied optics.

[46]  Leonard McMillan,et al.  Plenoptic Modeling: An Image-Based Rendering System , 2023 .

[47]  Jia Wu,et al.  A Correlation-Based Feature Weighting Filter for Naive Bayes , 2019, IEEE Transactions on Knowledge and Data Engineering.