Infrared and visible image fusion using dual discriminators generative adversarial networks with Wasserstein distance

Abstract Generative adversarial network (GAN) has shown great potential in infrared and visible image fusion. The existing GAN-based methods establish an adversarial game between generative image and source images to train the generator until the generative image contains enough meaningful information from source images. However, they only design one discriminator to force the fused result to complement gradient information from visible image, which may lose some detail information that existing in infrared image and omit some texture information that existing in visible image. To this end, we propose an end-to-end dual discriminators Wasserstein generative adversarial network, termed as D2WGAN, a framework that extends GAN to dual discriminators. In D2WGAN, the fused image can keep pixel intensity and details of infrared image by the first discriminator, and capture rich texture information of visible image by the second discriminator. In addition, to improve the performance of D2WGAN, we employ the GAN with Wasserstein distance. Moreover, in order to make the fused image keep more details from visible image in texture feature domain, we define a novel LBP (local binary pattern) loss. The extensive qualitative and quantitative experiments on public datasets demonstrate that D2WGAN can generate better results compared with the other state-of-the-art methods.

[1]  Liqiang Nie,et al.  Low-Rank Regularized Multi-Representation Learning for Fashion Compatibility Prediction , 2020, IEEE Transactions on Multimedia.

[2]  Léon Bottou,et al.  Wasserstein Generative Adversarial Networks , 2017, ICML.

[3]  Christian Ledig,et al.  Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[5]  Paul M. de Zeeuw,et al.  Fast saliency-aware multi-modality image fusion , 2013, Neurocomputing.

[6]  Paul S. Fisher,et al.  Image quality measures and their performance , 1995, IEEE Trans. Commun..

[7]  Shutao Li,et al.  Group-Sparse Representation With Dictionary Learning for Medical Image Denoising and Fusion , 2012, IEEE Transactions on Biomedical Engineering.

[8]  Aaron C. Courville,et al.  Improved Training of Wasserstein GANs , 2017, NIPS.

[9]  Meng Liu,et al.  Online Data Organizer: Micro-Video Categorization by Structure-Guided Multimodal Dictionary Learning , 2019, IEEE Transactions on Image Processing.

[10]  Xuanqin Mou,et al.  Low-Dose CT Image Denoising Using a Generative Adversarial Network With Wasserstein Distance and Perceptual Loss , 2017, IEEE Transactions on Medical Imaging.

[11]  Liqiang Nie,et al.  Neural Multimodal Cooperative Learning Toward Micro-Video Understanding , 2020, IEEE Transactions on Image Processing.

[12]  Jufeng Zhao,et al.  Fusion of visible and infrared images using global entropy and gradient constrained regularization , 2017 .

[13]  W. Kong,et al.  Adaptive fusion method of visible light and infrared images based on non-subsampled shearlet transform and fast non-negative matrix factorization , 2014 .

[14]  Jiayi Ma,et al.  Infrared and visible image fusion methods and applications: A survey , 2018, Inf. Fusion.

[15]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[16]  Bin Yang,et al.  Multi-focus image fusion and super-resolution with convolutional neural network , 2017, Int. J. Wavelets Multiresolution Inf. Process..

[17]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[18]  Jun Chen,et al.  Infrared and visible image fusion based on target-enhanced multiscale transform decomposition , 2020, Inf. Sci..

[19]  Edward H. Adelson,et al.  The Laplacian Pyramid as a Compact Image Code , 1983, IEEE Trans. Commun..

[20]  Jiayi Ma,et al.  Infrared and visible image fusion via gradient transfer and total variation minimization , 2016, Inf. Fusion.

[21]  S. Rajkumar,et al.  Infrared and Visible Image Fusion Using Entropy and Neuro-Fuzzy Concepts , 2014 .

[22]  Josef Kittler,et al.  Infrared and Visible Image Fusion using a Deep Learning Framework , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[23]  Qiang Zhang,et al.  Multifocus image fusion using the nonsubsampled contourlet transform , 2009, Signal Process..

[24]  Yi Chai,et al.  A novel dictionary learning approach for multi-modality medical image fusion , 2016, Neurocomputing.

[25]  Yu Liu,et al.  A general framework for image fusion based on multi-scale transform and sparse representation , 2015, Inf. Fusion.

[26]  Cedric Nishan Canagarajah,et al.  Pixel- and region-based image fusion with complex wavelets , 2007, Inf. Fusion.

[27]  Yi Liu,et al.  Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review , 2018, Inf. Fusion.

[28]  Dapeng Tao,et al.  Discriminative Dictionary Learning-Based Multiple Component Decomposition for Detail-Preserving Noisy Image Fusion , 2020, IEEE Transactions on Instrumentation and Measurement.

[29]  Qi Li,et al.  Infrared image enhancement through saliency feature analysis based on multi-scale decomposition , 2014 .

[30]  Meng Wang,et al.  Low-Rank Multi-View Embedding Learning for Micro-Video Popularity Prediction , 2018, IEEE Transactions on Knowledge and Data Engineering.

[31]  Luciano Alparone,et al.  Remote sensing image fusion using the curvelet transform , 2007, Inf. Fusion.

[32]  Matti Pietikäinen,et al.  Performance evaluation of texture measures with classification based on Kullback discrimination of distributions , 1994, Proceedings of 12th International Conference on Pattern Recognition.

[33]  Raymond Y. K. Lau,et al.  Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[34]  Alexander Toet,et al.  Image fusion by a ration of low-pass pyramid , 1989, Pattern Recognit. Lett..

[35]  Zhen Li,et al.  Coupled GAN With Relativistic Discriminators for Infrared and Visible Images Fusion , 2019, IEEE Sensors Journal.

[36]  Rabab Kreidieh Ward,et al.  Deep learning for pixel-level image fusion: Recent advances and future prospects , 2018, Inf. Fusion.

[37]  Wei Yu,et al.  Learning a Generative Model for Fusing Infrared and Visible Images via Conditional Generative Adversarial Network with Dual Discriminators , 2019, IJCAI.

[38]  Xinming Tang,et al.  IMAGE FUSION AND IMAGE QUALITY ASSESSMENT OF FUSED IMAGES , 2013 .

[39]  Haitao Yin,et al.  Sparse representation with learned multiscale dictionary for image fusion , 2015, Neurocomputing.

[40]  Junjun Jiang,et al.  FusionGAN: A generative adversarial network for infrared and visible image fusion , 2019, Inf. Fusion.