Learning to See Through Turbulent Water

Imaging through dynamic refractive media, such as looking into turbulent water, or through hot air, is challenging since light rays are bent by unknown amounts leading to complex geometric distortions. Inverting these distortions and recovering high quality images is an inherently ill-posed problem, leading previous works to require extra information such as high frame-rate video or a template image, which limits their applicability in practice. This paper proposes training a deep convolution neural network to undistort dynamic refractive effects using only a single image. The neural network is able to solve this ill-posed problem by learning image priors as well as distortion priors. Our network consists of two parts, a warping net to remove geometric distortion and a color predictor net to further refine the restoration. Adversarial loss is used to achieve better visual quality and help the network hallucinate missing and blurred information. To train our network, we collect a large training set of images distorted by a turbulent water surface. Unlike prior works on water undistortion, our method is trained end-to-end, only requires a single image and does not use a ground truth template at test time. Experiments show that by exploiting the structure of the problem, our network outperforms state-of-the-art deep image to image translation.

[1]  David W. Jacobs,et al.  WarpNet: Weakly Supervised Matching for Single-View Reconstruction , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  Alexei A. Efros,et al.  Generative Visual Manipulation on the Natural Image Manifold , 2016, ECCV.

[3]  Alexei A. Efros,et al.  Seeing through water , 2004, NIPS.

[4]  Yuandong Tian,et al.  Seeing through water: Image restoration using model-based tracking , 2009, 2009 IEEE 12th International Conference on Computer Vision.

[5]  Jitendra Malik,et al.  Learning to See by Moving , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[6]  Alexei A. Efros,et al.  Context Encoders: Feature Learning by Inpainting , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Frédo Durand,et al.  Refraction Wiggles for Measuring Fluid Depth and Velocity from Video , 2014, ECCV.

[8]  David A. Forsyth,et al.  Learning Large-Scale Automatic Image Colorization , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[9]  Yuandong Tian,et al.  Globally Optimal Estimation of Nonrigid Image Distortion , 2012, International Journal of Computer Vision.

[10]  Chuan Li,et al.  Precomputed Real-Time Texture Synthesis with Markovian Generative Adversarial Networks , 2016, ECCV.

[11]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[12]  Andrea Vedaldi,et al.  Instance Normalization: The Missing Ingredient for Fast Stylization , 2016, ArXiv.

[13]  Vishal M. Patel,et al.  Image De-Raining Using a Conditional Generative Adversarial Network , 2017, IEEE Transactions on Circuits and Systems for Video Technology.

[14]  Andrew Zisserman,et al.  Spatial Transformer Networks , 2015, NIPS.

[15]  Hiroshi Murase,et al.  Surface Shape Reconstruction of a Nonrigid Transport Object Using Refraction and Motion , 1992, IEEE Trans. Pattern Anal. Mach. Intell..

[16]  Aditya Deshpande,et al.  Learning Diverse Image Colorization , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Feng Zhou,et al.  Deep Deformation Network for Object Landmark Localization , 2016, ECCV.

[18]  Eraldo Ribeiro,et al.  Improved reconstruction of images distorted by water waves , 2006, VISAPP.

[19]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[20]  Qionghai Dai,et al.  Recovering Scene Geometry under Wavy Fluid via Distortion and Defocus Analysis , 2014, ECCV.

[21]  Alexei A. Efros,et al.  Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Ting-Chun Wang,et al.  Learning-based view synthesis for light field cameras , 2016, ACM Trans. Graph..

[23]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[24]  Zhen Wang,et al.  Multi-class Generative Adversarial Networks with the L2 Loss Function , 2016, ArXiv.

[25]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  Yuandong Tian,et al.  Theory and Practice of Hierarchical Data-driven Descent for Optimal Deformation Estimation , 2015, International Journal of Computer Vision.

[27]  Xiaoou Tang,et al.  Image Super-Resolution Using Deep Convolutional Networks , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[28]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[29]  Hiroshi Ishikawa,et al.  Let there be color! , 2016, ACM Trans. Graph..

[30]  Eraldo Ribeiro,et al.  Classification of Textures Distorted by WaterWaves , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[31]  Alexei A. Efros,et al.  Colorful Image Colorization , 2016, ECCV.

[32]  Andrew Lambert,et al.  Bispectral analysis and recovery of images distorted by a moving water surface. , 2010, Applied optics.

[33]  Yoav Y. Schechner,et al.  Triangulation in random refractive distortions , 2013, ICCP.

[34]  Mubarak Shah,et al.  A two-stage reconstruction approach for seeing through water , 2011, CVPR 2011.