DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion

In this paper, we proposed a new end-to-end model, termed as dual-discriminator conditional generative adversarial network (DDcGAN), for fusing infrared and visible images of different resolutions. Our method establishes an adversarial game between a generator and two discriminators. The generator aims to generate a real-like fused image based on a specifically designed content loss to fool the two discriminators, while the two discriminators aim to distinguish the structure differences between the fused image and two source images, respectively, in addition to the content loss. Consequently, the fused image is forced to simultaneously keep the thermal radiation in the infrared image and the texture details in the visible image. Moreover, to fuse source images of different resolutions, e.g., a low-resolution infrared image and a high-resolution visible image, our DDcGAN constrains the downsampled fused image to have similar property with the infrared image. This can avoid causing thermal radiation information blurring or visible texture detail loss, which typically happens in traditional methods. In addition, we also apply our DDcGAN to fusing multi-modality medical images of different resolutions, e.g., a low-resolution positron emission tomography image and a high-resolution magnetic resonance image. The qualitative and quantitative experiments on publicly available datasets demonstrate the superiority of our DDcGAN over the state-of-the-art, in terms of both visual effect and quantitative metrics. Our code is publicly available at https://github.com/jiayi-ma/DDcGAN.

[1]  Haixu Wang,et al.  Multimodal medical image fusion based on IHS and PCA , 2010 .

[2]  Yi Liu,et al.  Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review , 2018, Inf. Fusion.

[3]  Junjun Jiang,et al.  FusionGAN: A generative adversarial network for infrared and visible image fusion , 2019, Inf. Fusion.

[4]  Vikrant Bhateja,et al.  Multimodal Medical Image Sensor Fusion Framework Using Cascade of Wavelet and Contourlet Transform Domains , 2015, IEEE Sensors Journal.

[5]  Xuelong Li,et al.  Image Super-Resolution With Sparse Neighbor Embedding , 2012, IEEE Transactions on Image Processing.

[6]  Ian J. Goodfellow,et al.  NIPS 2016 Tutorial: Generative Adversarial Networks , 2016, ArXiv.

[7]  Ying Han,et al.  Structure-aware image fusion , 2018, Optik.

[8]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[9]  Ali Mohammadzadeh,et al.  Developing a Spectral-Based Strategy for Urban Object Detection From Airborne Hyperspectral TIR and Visible Data , 2016, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[10]  Josef Kittler,et al.  Infrared and Visible Image Fusion using a Deep Learning Framework , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[11]  Humberto Bustince,et al.  Self-adapting weighted operators for multiscale gradient fusion , 2018, Inf. Fusion.

[12]  Davide Cozzolino,et al.  Pansharpening by Convolutional Neural Networks , 2016, Remote. Sens..

[13]  Marc Teboulle,et al.  Fast Gradient-Based Algorithms for Constrained Total Variation Image Denoising and Deblurring Problems , 2009, IEEE Transactions on Image Processing.

[14]  Hui Li,et al.  DenseFuse: A Fusion Approach to Infrared and Visible Images , 2018, IEEE Transactions on Image Processing.

[15]  Zhen Li,et al.  Coupled GAN With Relativistic Discriminators for Infrared and Visible Images Fusion , 2019, IEEE Sensors Journal.

[16]  Pragati Upadhyay,et al.  PIXEL-LEVEL IMAGE FUSION USINGBROVEY TRANSFORME AND WAVELETTRANSFORM , 2013 .

[17]  Yu Liu,et al.  Simultaneous image fusion and denoising with adaptive sparse representation , 2015, IET Image Process..

[18]  Shaowen Yao,et al.  A survey of infrared and visual image fusion methods , 2017 .

[19]  Jun Huang,et al.  Fusing Infrared and Visible Images of Different Resolutions via Total Variation Model , 2018, Sensors.

[20]  Rabab Kreidieh Ward,et al.  Deep learning for pixel-level image fusion: Recent advances and future prospects , 2018, Inf. Fusion.

[21]  Wei Yu,et al.  Learning a Generative Model for Fusing Infrared and Visible Images via Conditional Generative Adversarial Network with Dual Discriminators , 2019, IJCAI.

[22]  Garry E Gold,et al.  Potential of PET-MRI for imaging of non-oncologic musculoskeletal disease. , 2016, Quantitative imaging in medicine and surgery.

[23]  Rabab Kreidieh Ward,et al.  Image Fusion With Convolutional Sparse Representation , 2016, IEEE Signal Processing Letters.

[24]  Li Yan,et al.  A fusion algorithm for infrared and visible images based on adaptive dual-channel unit-linking PCNN in NSCT domain , 2015 .

[25]  Vps Naidu,et al.  Hybrid DDCT-PCA based multi sensor image fusion , 2014 .

[26]  Jun Chen,et al.  Infrared and visible image fusion using total variation model , 2016, Neurocomputing.

[27]  Qiang Guo,et al.  An Adaptive Fusion Algorithm for Visible and Infrared Videos Based on Entropy and the Cumulative Distribution of Gray Levels , 2017, IEEE Transactions on Multimedia.

[28]  Jie Zhan,et al.  Carbon quantum dots/hydrogenated TiO2 nanobelt heterostructures and their broad spectrum photocatalytic properties under UV, visible, and near-infrared irradiation , 2015 .

[29]  Tsubasa Saito,et al.  Middle infrared (wavelength range: 8 μm-14 μm) 2-dimensional spectroscopy (total weight with electrical controller: 1.7 kg, total cost: less than 10,000 USD) so-called hyperspectral camera for unmanned air vehicles like drones , 2016, SPIE Defense + Security.

[30]  Vishal M. Patel,et al.  Image De-Raining Using a Conditional Generative Adversarial Network , 2017, IEEE Transactions on Circuits and Systems for Video Technology.

[31]  Yu Liu,et al.  A medical image fusion method based on convolutional neural networks , 2017, 2017 20th International Conference on Information Fusion (Fusion).

[32]  Alan L. Yuille,et al.  Estimation of 3D Category-Specific Object Structure: Symmetry, Manhattan and/or Multiple Images , 2019, International Journal of Computer Vision.

[33]  Yang Chao,et al.  Efficient image fusion with approximate sparse representation , 2016, Int. J. Wavelets Multiresolution Inf. Process..

[34]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[35]  R. Venkatesh Babu,et al.  DeepFuse: A Deep Unsupervised Approach for Exposure Fusion with Extreme Exposure Image Pairs , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[36]  Graham W. Taylor,et al.  Adaptive deconvolutional networks for mid and high level feature learning , 2011, 2011 International Conference on Computer Vision.

[37]  Hua Zong,et al.  Infrared and visible image fusion based on visual saliency map and weighted least square optimization , 2017 .

[38]  J. Wesley Roberts,et al.  Assessment of image fusion procedures using entropy, image quality, and multispectral classification , 2008 .

[39]  Quan Wang,et al.  Infrared and visible image fusion based on target extraction in the nonsubsampled contourlet transform domain , 2017 .

[40]  Wei Liu,et al.  A novel infrared and visible image fusion algorithm based on shift-invariant dual-tree complex shearlet transform and sparse representation , 2017, Neurocomputing.

[41]  Simon Osindero,et al.  Conditional Generative Adversarial Nets , 2014, ArXiv.

[42]  Shutao Li,et al.  Image Fusion With Guided Filtering , 2013, IEEE Transactions on Image Processing.

[43]  Ashish Khare,et al.  Fusion of multimodal medical images using Daubechies complex wavelet transform - A multiresolution approach , 2014, Inf. Fusion.

[44]  Jiayi Ma,et al.  Infrared and visible image fusion methods and applications: A survey , 2018, Inf. Fusion.

[45]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[46]  Sunil Agrawal,et al.  From Multi-Scale Decomposition to Non-Multi-Scale Decomposition Methods: A Comprehensive Survey of Image Fusion Techniques and Its Applications , 2017, IEEE Access.

[47]  Sabalan Daneshvar,et al.  MRI and PET image fusion by combining IHS and retina-inspired models , 2010, Inf. Fusion.

[48]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[49]  Gang Liu,et al.  Multi-sensor image fusion based on fourth order partial differential equations , 2017, 2017 20th International Conference on Information Fusion (Fusion).

[50]  Jiayi Ma,et al.  Infrared and visible image fusion via gradient transfer and total variation minimization , 2016, Inf. Fusion.

[51]  Vinod Kumar,et al.  Performance Evaluation of Color Models in the Fusion of Functional and Anatomical Images , 2016, Journal of Medical Systems.

[52]  Paul S. Fisher,et al.  Image quality measures and their performance , 1995, IEEE Trans. Commun..

[53]  Li Bai,et al.  Implementation of high‐order variational models made easy for image processing , 2016 .

[54]  Wei Yu,et al.  Infrared and visible image fusion via detail preserving adversarial learning , 2020, Inf. Fusion.

[55]  B. K. Shreyamsha Kumar,et al.  Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform , 2013, Signal Image Video Process..

[56]  Yu Han,et al.  A new image fusion performance metric based on visual information fidelity , 2013, Inf. Fusion.

[57]  Sun Li,et al.  Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters , 2016, Inf. Fusion.

[58]  Robert Matthews,et al.  Clinical Utility of Positron Emission Tomography Magnetic Resonance Imaging (PET-MRI) in Gastrointestinal Cancers , 2016, Diagnostics.