Different Input Resolutions and Arbitrary Output Resolution: A Meta Learning-Based Deep Framework for Infrared and Visible Image Fusion

Infrared and visible image fusion has gained ever-increasing attention in recent years due to its great significance in a variety of vision-based applications. However, existing fusion methods suffer from some limitations in terms of the spatial resolutions of both input source images and output fused image, which prevents their practical usage to a great extent. In this paper, we propose a meta learning-based deep framework for the fusion of infrared and visible images. Unlike most existing methods, the proposed framework can accept the source images of different resolutions and generate the fused image of arbitrary resolution just with a single learned model. In the proposed framework, the features of each source image are first extracted by a convolutional network and upscaled by a meta-upscale module with an arbitrary appropriate factor according to practical requirements. Then, a dual attention mechanism-based feature fusion module is developed to combine features from different source images. Finally, a residual compensation module, which can be iteratively adopted in the proposed framework, is designed to enhance the capability of our method in detail extraction. In addition, the loss function is formulated in a multi-task learning manner via simultaneous fusion and super-resolution, aiming to improve the effect of feature learning. And, a new contrast loss inspired by a perceptual contrast enhancement approach is proposed to further improve the contrast of the fused image. Extensive experiments on widely-used fusion datasets demonstrate the effectiveness and superiority of the proposed method. The code of the proposed method is publicly available at https://github.com/yuliu316316/MetaLearning-Fusion.

[1]  Suyash P. Awate,et al.  Robust Super-Resolution Gan, with Manifold-Based and Perception Loss , 2019, 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019).

[2]  Seoung Wug Oh,et al.  Deep Video Super-Resolution Network Using Dynamic Upsampling Filters Without Explicit Motion Compensation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[3]  David Bull,et al.  Image fusion metric based on mutual information and Tsallis entropy , 2006 .

[4]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[5]  Jiayi Ma,et al.  Infrared and visible image fusion via gradient transfer and total variation minimization , 2016, Inf. Fusion.

[6]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[7]  Yufeng Zheng,et al.  A new metric based on extended spatial frequency and its application to DWT based fusion algorithms , 2007, Inf. Fusion.

[8]  Yafei Zhang,et al.  Infrared and visible image fusion scheme based on NSCT and low-level visual features , 2016 .

[9]  Shutao Li,et al.  Simultaneous image fusion and super-resolution using sparse representation , 2013, Inf. Fusion.

[10]  Gregory Shakhnarovich,et al.  Deep Back-Projection Networks for Super-Resolution , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[11]  Yi Chai,et al.  A novel multi-modality image fusion method based on image decomposition and sparse representation , 2017, Inf. Sci..

[12]  L. Wald,et al.  Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images , 1997 .

[13]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[14]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Tieniu Tan,et al.  Meta-SR: A Magnification-Arbitrary Network for Super-Resolution , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  Nam Ik Cho,et al.  Meta-Transfer Learning for Zero-Shot Super-Resolution , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Alessandro Rizzi,et al.  Perceptual Color Correction Through Variational Techniques , 2007, IEEE Transactions on Image Processing.

[18]  Yu Liu,et al.  IFCNN: A general image fusion framework based on convolutional neural network , 2020, Inf. Fusion.

[19]  Jiayi Ma,et al.  Infrared and visible image fusion methods and applications: A survey , 2018, Inf. Fusion.

[20]  Qi Li,et al.  Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition , 2015 .

[21]  Yu Liu,et al.  Multi-focus image fusion with a deep convolutional neural network , 2017, Inf. Fusion.

[22]  R. Venkatesh Babu,et al.  DeepFuse: A Deep Unsupervised Approach for Exposure Fusion with Extreme Exposure Image Pairs , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[23]  Zhengtao Yu,et al.  Fractional differential and variational method for image fusion and super-resolution , 2016, Neurocomputing.

[24]  Richa Singh,et al.  Integrated multilevel image fusion and match score fusion of visible and infrared face images for robust face recognition , 2008, Pattern Recognit..

[25]  Hui Li,et al.  DenseFuse: A Fusion Approach to Infrared and Visible Images , 2018, IEEE Transactions on Image Processing.

[26]  Yu Qiao,et al.  ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks , 2018, ECCV Workshops.

[27]  Mahboob Iqbal,et al.  Unification of image fusion and super-resolution using jointly trained dictionaries and local information contents , 2012 .

[28]  Long Wang,et al.  Multisensor video fusion based on higher order singular value decomposition , 2015, Inf. Fusion.

[29]  Yafei Zhang,et al.  Analysis-synthesis dictionary pair learning and patch saliency measure for image fusion , 2020, Signal Process..

[30]  Feng Wu,et al.  DRPL: Deep Regression Pair Learning for Multi-Focus Image Fusion , 2020, IEEE Transactions on Image Processing.

[31]  Gemma Piella,et al.  Image Fusion for Enhanced Visualization: A Variational Approach , 2009, International Journal of Computer Vision.

[32]  Yunhong Wang,et al.  Receptive Field Block Net for Accurate and Fast Object Detection , 2017, ECCV.

[33]  Junjun Jiang,et al.  FusionGAN: A generative adversarial network for infrared and visible image fusion , 2019, Inf. Fusion.

[34]  Yu Liu,et al.  A general framework for image fusion based on multi-scale transform and sparse representation , 2015, Inf. Fusion.

[35]  Yongmei Cheng,et al.  Method of visual and infrared fusion for moving object detection. , 2013, Optics letters.

[36]  Shutao Li,et al.  Pixel-level image fusion: A survey of the state of the art , 2017, Inf. Fusion.

[37]  Rabab Kreidieh Ward,et al.  Deep learning for pixel-level image fusion: Recent advances and future prospects , 2018, Inf. Fusion.

[38]  Yafei Zhang,et al.  Joint Framework for Image Fusion and Super-Resolution via Multicomponent Analysis and Residual Compensation , 2019, IEEE Access.

[39]  Xiuqing Wu,et al.  A novel similarity based quality metric for image fusion , 2008, 2008 International Conference on Audio, Language and Image Processing.

[40]  Shaowen Yao,et al.  A survey of infrared and visual image fusion methods , 2017 .

[41]  Rabab Kreidieh Ward,et al.  Image Fusion With Convolutional Sparse Representation , 2016, IEEE Signal Processing Letters.

[42]  Xiao-Ping Zhang,et al.  DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion , 2020, IEEE Transactions on Image Processing.

[43]  Kwanghoon Sohn,et al.  Unsupervised Deep Image Fusion With Structure Tensor Representations , 2020, IEEE Transactions on Image Processing.

[44]  Shutao Li,et al.  Pixel-level image fusion with simultaneous orthogonal matching pursuit , 2012, Inf. Fusion.

[45]  Gwanggil Jeon,et al.  A Symmetric Encoder-Decoder with Residual Block for Infrared and Visible Image Fusion , 2019, ArXiv.

[46]  Yu Liu,et al.  Infrared and visible image fusion with convolutional neural networks , 2017, Int. J. Wavelets Multiresolution Inf. Process..

[47]  Sabine Süsstrunk,et al.  Fast and Efficient Zero-Learning Image Fusion , 2019, ArXiv.

[48]  M. Hossny,et al.  Comments on 'Information measure for performance of image fusion' , 2008 .

[49]  Jun Chen,et al.  Infrared and visible image fusion based on target-enhanced multiscale transform decomposition , 2020, Inf. Sci..