SMFuse: Multi-Focus Image Fusion Via Self-Supervised Mask-Optimization

In this paper, a novel self-supervised mask-optimization model, termed as SMFuse, is proposed for multi-focus image fusion. In our model, given two source images, a fully end-to-end Mask-Generator is trained to directly generate the binary mask without requiring any patch operation or postprocessing through self-supervised learning. On the one hand, based on the principle of repeated blur, we design a Guided-Block with guided filter to obtain an initial binary mask from source images, narrowing the solution domain and speeding up the convergence of the binary mask generation, which is constrained by a map loss. On the other hand, as the focused regions in source images show richer texture details than the defocused ones, i.e., larger gradients, we also design a max-gradient loss between the fused image and source images as a follow-up optimization operation to ensure the fused image to be all-in-focus, forcing our model to learn a more accurate binary mask. Extensive experimental results conducted on two publicly available datasets substantiate the effectiveness and superiority of our SMFuse compared with the current state-of-the-art. Our code is publicly available online.11[Online]. Available: https://github.com/jiayi-ma/SMFuse.

[1]  Yi Liu,et al.  Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review , 2018, Inf. Fusion.

[2]  Lei Zhang,et al.  Learning a Deep Single Image Contrast Enhancer from Multi-Exposure Images , 2018, IEEE Transactions on Image Processing.

[3]  Lei Xiong,et al.  VIF-Net: An Unsupervised Framework for Infrared and Visible Image Fusion , 2020, IEEE Transactions on Computational Imaging.

[4]  Qi Tian,et al.  Codebook Guided Feature-Preserving for Recognition-Oriented Image Retargeting , 2017, IEEE Transactions on Image Processing.

[5]  Feng Wu,et al.  DRPL: Deep Regression Pair Learning for Multi-Focus Image Fusion , 2020, IEEE Transactions on Image Processing.

[6]  Shutao Li,et al.  Image Fusion With Guided Filtering , 2013, IEEE Transactions on Image Processing.

[7]  Jun Huang,et al.  A Deep Model for Multi-Focus Image Fusion Based on Gradients and Connected Regions , 2020, IEEE Access.

[8]  Michael I. Jordan,et al.  Unsupervised Domain Adaptation with Residual Transfer Networks , 2016, NIPS.

[9]  Philipp Hennig,et al.  Dissecting Adam: The Sign, Magnitude and Variance of Stochastic Gradients , 2017, ICML.

[10]  Jiayi Ma,et al.  Variational Pansharpening by Exploiting Cartoon-Texture Similarities , 2021, IEEE Transactions on Geoscience and Remote Sensing.

[11]  Pietro Perona,et al.  Microsoft COCO: Common Objects in Context , 2014, ECCV.

[12]  Junjun Jiang,et al.  FusionGAN: A generative adversarial network for infrared and visible image fusion , 2019, Inf. Fusion.

[13]  Hui Li,et al.  DenseFuse: A Fusion Approach to Infrared and Visible Images , 2018, IEEE Transactions on Image Processing.

[14]  Xi Chen,et al.  PixelCNN++: Improving the PixelCNN with Discretized Logistic Mixture Likelihood and Other Modifications , 2017, ICLR.

[15]  Hao Zhang,et al.  Rethinking the Image Fusion: A Fast Unified Image Fusion Network based on Proportional Maintenance of Gradient and Intensity , 2020, AAAI.

[16]  Yu Liu,et al.  Multi-focus image fusion with dense SIFT , 2015, Inf. Fusion.

[17]  Fan Fan,et al.  A generative adversarial network with adaptive constraints for multi-focus image fusion , 2020, Neural Computing and Applications.

[18]  Yong Yang,et al.  Multi-focus Image Fusion Using an Effective Discrete Wavelet Transform Based Algorithm , 2014 .

[19]  Milind Tambe,et al.  End to end learning and optimization on graphs , 2019, NeurIPS.

[20]  Han Xu,et al.  MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion , 2021, Inf. Fusion.

[21]  Xiaohua Qiu,et al.  Guided filter-based multi-focus image fusion through focus region detection , 2019, Signal Process. Image Commun..

[22]  Roberto Manduchi,et al.  Bilateral filtering for gray and color images , 1998, Sixth International Conference on Computer Vision (IEEE Cat. No.98CH36271).

[23]  Yu Liu,et al.  Multi-focus image fusion with a deep convolutional neural network , 2017, Inf. Fusion.

[24]  Jian Sun,et al.  Guided Image Filtering , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Xiaojun Wu,et al.  Multi-focus image fusion using HOSVD and edge intensity , 2017, J. Vis. Commun. Image Represent..

[26]  Panajotis Agathoklis,et al.  Multi-Exposure and Multi-Focus Image Fusion in Gradient Domain , 2016, J. Circuits Syst. Comput..

[27]  Yong Yang,et al.  Multilevel Features Convolutional Neural Network for Multifocus Image Fusion , 2019, IEEE Transactions on Computational Imaging.

[28]  Tania Stathaki,et al.  Image Fusion: Algorithms and Applications , 2008 .

[29]  Paul S. Fisher,et al.  Image quality measures and their performance , 1995, IEEE Trans. Commun..

[30]  Vladimir Petrovic,et al.  Objective image fusion performance measure , 2000 .

[31]  Xiaojie Guo,et al.  U2Fusion: A Unified Unsupervised Image Fusion Network , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  Naoto Yokoya,et al.  CoSpace: Common Subspace Learning From Hyperspectral-Multispectral Correspondences , 2018, IEEE Transactions on Geoscience and Remote Sensing.

[33]  Xiao-Ping Zhang,et al.  DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion , 2020, IEEE Transactions on Image Processing.

[34]  Kwanghoon Sohn,et al.  Unsupervised Deep Image Fusion With Structure Tensor Representations , 2020, IEEE Transactions on Image Processing.

[35]  Raymond Y. K. Lau,et al.  Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[36]  Huchuan Lu,et al.  Multi-Focus Image Fusion With a Natural Enhancement via a Joint Multi-Level Deeply Supervised Convolutional Neural Network , 2019, IEEE Transactions on Circuits and Systems for Video Technology.

[37]  Xiaojuan Ban,et al.  SESF-Fuse: An Unsupervised Deep Model for Multi-Focus Image Fusion , 2019, ArXiv.

[38]  Yi Chai,et al.  Multi-focus image fusion based on nonsubsampled contourlet transform and focused regions detection , 2013 .

[39]  K. Batenburg,et al.  Noise2Inverse: Self-Supervised Deep Convolutional Denoising for Tomography , 2020, IEEE Transactions on Computational Imaging.

[40]  Yoshua Bengio,et al.  BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.

[41]  Jiayi Ma,et al.  Infrared and visible image fusion methods and applications: A survey , 2018, Inf. Fusion.

[42]  Charles A. Bouman,et al.  A Framework for Dynamic Image Sampling Based on Supervised Learning , 2017, IEEE Transactions on Computational Imaging.

[43]  Yibo Chen,et al.  Robust Multi-Focus Image Fusion Using Edge Model and Multi-Matting , 2018, IEEE Transactions on Image Processing.

[44]  Naoto Yokoya,et al.  Learnable manifold alignment (LeMA): A semi-supervised cross-modality learning framework for land cover and land use classification , 2019, ISPRS journal of photogrammetry and remote sensing : official publication of the International Society for Photogrammetry and Remote Sensing.