Zero-Shot Learning of Continuous 3D Refractive Index Maps from Discrete Intensity-Only Measurements

Intensity diffraction tomography (IDT) refers to a class of optical microscopy techniques for imaging the 3D refractive index (RI) distribution of a sample from a set of 2D intensity-only measurements. The reconstruction of artifact-free RI maps is a fundamental challenge in IDT due to the loss of phase information and the missing cone problem. Neural fields (NF) has recently emerged as a new deep learning (DL) paradigm for learning continuous representations of complex 3D scenes without external training datasets. We present DeCAF as the first NF-based IDT method that can learn a high-quality continuous representation of a RI volume directly from its intensity-only and limited-angle measurements. We show on three different IDT modalities and multiple biological samples that DeCAF can generate high-contrast and artifact-free RI maps. Refractive index (RI) is an optical property that describes the interaction between light and matter within a sample. The real part of RI characterizes the phase velocity while its imaginary part characterizes the attenuation. RI can thus serve as an endogenous source of optical contrast for imaging samples without staining or labeling them using external contrast agents. By quantitatively characterizing the three-dimensional (3D) distribution of the RI, one can visualize cellular or subcellular structures useful in morphogenesis [1], oncology [2], cellular pathophysiology [3], biochemistry [4], and beyond (see the review papers [5,6]). Intensity diffraction tomography (IDT) is a recent technique for producing 3D RI maps of a sample by measuring the light it diffracts. In the standard IDT setup, a sample is illuminated multiple times from different angles, and a set of two-dimensional (2D) intensity projections are captured by the camera (see Figure 1(a)). A tomographic image reconstruction algorithm is then used to computationally reconstruct the desired 3D RI distribution from the set of 2D measurements. Unlike traditional optical diffraction tomography (ODT) that uses interferometry to record the complex-valued light fields [7–9], IDT only measures the squared amplitude of the scattered light, leading to an easy setup on standard transmission optical microscopes with inexpensive hardware modifications. Such flexibility has spurred different IDT variants integrating object scanning [10,11], angled illumination [12–15], pupil engineering [16,17], and multiple scattering [18,19]. Setups achieving high resolution [18] and fast acquisition [20] have also been reported. 1 ar X iv :2 11 2. 00 00 2v 1 [ ee ss .I V ] 2 7 N ov 2 02 1 Despite the rich literature on IDT, image reconstruction remains a fundamental challenge. The first issue is that the phase of the scattered light field is missing from the measurements, resulting in a nonlinear measurement system that is not characterizable by the classical linear Fourier diffraction theory [21]. This rules out the usage of the standard filtered-backprojection methods and calls for advanced computational algorithms. The second issue is the well-known missing cone problem, causing elongation of the object along the optical axis (z dimension) and hence reduction of the axial resolution. The missing cone problem is due to the limited-angle tomographic setup, where illuminations can come only from one side of the sample plane with a limited range for angle variation (less than about 40◦ in our setups). This leads to an incomplete coverage of the 3D Fourier spectra with a cone-shape missing region in the axial direction. These missing phase and missing cone problems make image reconstruction in IDT a severely ill-posed inverse problem. Regularization methods are widely-used for mitigating ill-posed nature of many inverse problems. These methods are based on minimizing a cost function consisting of a data-fidelity term and a regularization term, where the former uses a physical-model to quantify the mismatch between the predicted and acquired measurements, while the latter promotes solutions that are consistent with a priori knowledge on the sample. For example, the least-squares loss and Tikhonov regularizer (`2penalty) are widely-used for obtaining closed-form solutions to inverse problems [14]. The work on plug-and-play priors has generalized the notion of image priors to implicit regularizers characterized by image denoisers [22–25]. Recently, deep learning (DL) has emerged as a powerful paradigm for image reconstruction. A traditional DL reconstruction is based on training a convolutional neural network (CNN) over a large dataset to learn a mapping from low-quality images to their high-quality counterparts. The state-of-the-art performance of such methods has been demonstrated in X-ray computed tomography [26,27], magnetic resonance imaging [28,29], optical tomography [30,31], and seismic imaging [32] (see the reviews [33–35]). While DL has significantly improved image reconstruction in many modalities, traditional DL methods are impractical for image reconstruction in IDT where it is difficult to acquire high-quality ground-truth RI maps in experiments. Although a physics-based simulator has been proposed to generate datasets for training IDT artifact-suppressing CNNs, the results are still limited by the mismatch between the simulation and experiments [36]. Neural fields (NF) is a recent DL paradigm that has gained popularity in computer vision and graphics for representing and rendering 3D scenes using coordinate-based deep neural networks [37, 38]. It is worth mentioning that while NF was suggested to be the most appropriate term [39,40], this idea currently goes by various names in the vision/graphics literature, including neural coordinatebased representations or neural implicit models. It has been shown that NF can learn a high-quality representation of a complex scene from a sparse set of data without any external training dataset. Motivated by this property, we propose Deep Continuous Artifact-free RI Field (DeCAF) as a first NF-based IDT method for learning a high-quality continuous 3D RI map from intensity-only and limited-angle measurements without any external training dataset of ground-truth RI maps. Figure 1 provides a conceptual illustration of DeCAF. The key features of DeCAF are as follows: • The central component of DeCAF is a multilayer perceptron (MLP), which is a fully-connected (non-convolutional) deep network, for learning a function that maps 3D coordinates (x, y, z) to the corresponding RI values. The trained MLP thus provides a continuous neural representation of the RI map. The RI value at any spatial location can be retrieved by querying the trained MLP with the corresponding coordinate. By decoupling representation from an explicit voxel grid, DeCAF can efficiently store and process large 3D volumes. • DeCAF is a zero-shot method, meaning that it does not require training using an external 2

[1]  L. Waller,et al.  Multi-layer Born multiple-scattering model for 3D phase microscopy , 2020 .

[2]  Dong Liang,et al.  Deep Magnetic Resonance Image Reconstruction: Inverse Problems Meet Neural Networks , 2020, IEEE Signal Processing Magazine.

[3]  Gordon Wetzstein,et al.  Scene Representation Networks: Continuous 3D-Structure-Aware Neural Scene Representations , 2019, NeurIPS.

[4]  L. Tian,et al.  3D intensity and phase imaging from light field measurements in an LED array microscope , 2015 .

[5]  Alex Matlock,et al.  High-speed in vitro intensity diffraction tomography , 2019, SPIE/COS Photonics Asia.

[6]  Chao Zuo,et al.  Three-dimensional tomographic microscopy technique with multi-frequency combination with partially coherent illuminations. , 2018, Biomedical optics express.

[7]  Gordon Wetzstein,et al.  Implicit Neural Representations with Periodic Activation Functions , 2020, NeurIPS.

[8]  Lei Zhang,et al.  Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising , 2016, IEEE Transactions on Image Processing.

[9]  Laura Waller,et al.  3D differential phase contrast microscopy , 2016, SPIE BiOS.

[10]  Lei Tian,et al.  Deep speckle correlation: a deep learning approach toward scalable imaging through scattering media , 2018, Optica.

[11]  Jonathan T. Barron,et al.  Deformable Neural Radiance Fields , 2020, ArXiv.

[12]  Michael Unser,et al.  Learning approach to optical tomography , 2015, 1502.01914.

[13]  Ulugbek Kamilov,et al.  Efficient and accurate inversion of multiple scattering with deep learning , 2018, Optics express.

[14]  Brendt Wohlberg,et al.  CoIL: Coordinate-Based Internal Learning for Tomographic Imaging , 2021, IEEE Transactions on Computational Imaging.

[15]  YongKeun Park,et al.  Measurements of three-dimensional refractive index tomography and membrane deformability of live erythrocytes from Pelophylax nigromaculatus , 2017, Scientific Reports.

[16]  E. Wolf Three-dimensional structure determination of semi-transparent objects from holographic data , 1969 .

[17]  Greg Gbur,et al.  Diffraction tomography without phase information. , 2002, Optics letters.

[18]  Ronald Clark,et al.  TermiNeRF: Ray Termination Prediction for Efficient Neural Rendering , 2021, 2021 International Conference on 3D Vision (3DV).

[19]  Jonathan T. Barron,et al.  Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains , 2020, NeurIPS.

[20]  Richard A. Newcombe,et al.  DeepSDF: Learning Continuous Signed Distance Functions for Shape Representation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[21]  Jitendra Malik,et al.  A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics , 2001, Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001.

[22]  Supasorn Suwajanakorn,et al.  NeX: Real-time View Synthesis with Neural Basis Expansion , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  J. Rogers,et al.  Spatial light interference microscopy (SLIM) , 2010, IEEE Photonic Society 24th Annual Meeting.

[24]  Jong Chul Ye,et al.  Deep learning for tomographic image reconstruction , 2020, Nature Machine Intelligence.

[25]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[26]  Lei Tian,et al.  Physical model simulator-trained neural network for computational 3D phase imaging of multiple-scattering samples , 2021, ArXiv.

[27]  Geoffrey E. Hinton,et al.  How to Represent Part-Whole Hierarchies in a Neural Network , 2021, Neural Computation.

[28]  Hujun Bao,et al.  Neural Body: Implicit Neural Representations with Structured Latent Codes for Novel View Synthesis of Dynamic Humans , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  Wangmeng Zuo,et al.  Learning Deep CNN Denoiser Prior for Image Restoration , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[30]  C. Depeursinge,et al.  Quantitative phase imaging in biomedicine , 2012, 2012 Conference on Lasers and Electro-Optics (CLEO).

[31]  Lei Tian,et al.  High-throughput intensity diffraction tomography with a computational microscope. , 2018, Biomedical optics express.

[32]  Matthew Tancik,et al.  pixelNeRF: Neural Radiance Fields from One or Few Images , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[33]  Brendt Wohlberg,et al.  Provable Convergence of Plug-and-Play Priors With MMSE Denoisers , 2020, IEEE Signal Processing Letters.

[34]  Hyojin Kim,et al.  Dynamic CT Reconstruction from Limited Views with Implicit Neural Representations and Parametric Motion Fields , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[35]  Charles A. Bouman,et al.  Plug-and-Play Priors for Bright Field Electron Tomography and Sparse Interpolation , 2015, IEEE Transactions on Computational Imaging.

[36]  Jong Chul Ye,et al.  A deep convolutional neural network using directional wavelets for low‐dose X‐ray CT reconstruction , 2016, Medical physics.

[37]  Tan H. Nguyen,et al.  Gradient light interference microscopy for 3D imaging of unlabeled specimens , 2017, Nature Communications.

[38]  Di Jin,et al.  Tomographic phase microscopy: principles and applications in bioimaging [Invited]. , 2017, Journal of the Optical Society of America. B, Optical physics.

[39]  Tomaso Poggio,et al.  Deep Learning for Seismic Inverse Problems: Toward the Acceleration of Geophysical Analysis Workflows , 2021, IEEE Signal Processing Magazine.

[40]  Stanley H. Chan,et al.  Plug-and-Play ADMM for Image Restoration: Fixed-Point Convergence and Applications , 2016, IEEE Transactions on Computational Imaging.

[41]  Yongkeun Park,et al.  Refractive index maps and membrane dynamics of human red blood cells parasitized by Plasmodium falciparum , 2008, Proceedings of the National Academy of Sciences.

[42]  Mathews Jacob,et al.  MoDL: Model-Based Deep Learning Architecture for Inverse Problems , 2017, IEEE Transactions on Medical Imaging.

[43]  Michael Unser,et al.  Deep Convolutional Neural Network for Inverse Problems in Imaging , 2016, IEEE Transactions on Image Processing.

[44]  Kenneth M. Yamada,et al.  Modeling Tissue Morphogenesis and Cancer in 3D , 2007, Cell.

[45]  Charles A. Bouman,et al.  Plug-and-Play Methods for Magnetic Resonance Imaging: Using Denoisers for Image Recovery , 2019, IEEE Signal Processing Magazine.

[46]  Kyoohyun Kim,et al.  Three-dimensional label-free imaging and quantification of lipid droplets in live hepatocytes , 2016, Scientific Reports.

[47]  Jonathan T. Barron,et al.  NeRF in the Wild: Neural Radiance Fields for Unconstrained Photo Collections , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[48]  Noah Snavely,et al.  Neural Scene Flow Fields for Space-Time View Synthesis of Dynamic Scenes , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[49]  Brendt Wohlberg,et al.  Plug-and-Play priors for model based reconstruction , 2013, 2013 IEEE Global Conference on Signal and Information Processing.

[50]  Yu Sun,et al.  Async-RED: A Provably Convergent Asynchronous Block Parallel Stochastic Method using Deep Denoising Priors , 2020, ICLR.

[51]  Pratul P. Srinivasan,et al.  NeRF , 2020, ECCV.

[52]  Brendt Wohlberg,et al.  Scalable Plug-and-Play ADMM with Convergence Guarantees , 2020, ArXiv.

[53]  Alex Matlock,et al.  High-throughput, volumetric quantitative phase imaging with multiplexed intensity diffraction tomography. , 2019, Biomedical optics express.

[54]  M. Glas,et al.  Principles of Computerized Tomographic Imaging , 2000 .

[55]  T. Gaylord,et al.  Three-dimensional quantitative phase imaging via tomographic deconvolution phase microscopy. , 2015, Applied optics.

[56]  Jonathan T. Barron,et al.  NeRV: Neural Reflectance and Visibility Fields for Relighting and View Synthesis , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[57]  Yongjin Sung,et al.  Multiple Phases of Chondrocyte Enlargement Underlie Differences in Skeletal Proportions , 2013, Nature.

[58]  Jiaming Liu,et al.  SIMBA: Scalable Inversion in Optical Tomography Using Deep Denoising Priors , 2019, IEEE Journal of Selected Topics in Signal Processing.

[59]  C. Fang-Yen,et al.  Optical diffraction tomography for high resolution live cell imaging. , 2009, Optics express.

[60]  Laura Waller,et al.  High-resolution 3D refractive index microscopy of multiple-scattering samples from intensity images. , 2019, Optica.

[61]  Zhongping Zhang,et al.  Data-Driven Seismic Waveform Inversion: A Study on the Robustness and Generalization , 2018, IEEE Transactions on Geoscience and Remote Sensing.