Image fusion based on generative adversarial network consistent with perception

Abstract Deep learning is a rapidly developing approach in the field of infrared and visible image fusion. In this context, the use of dense blocks in deep networks significantly improves the utilization of shallow information, and the combination of the Generative Adversarial Network (GAN) also improves the fusion performance of two source images. We propose a new method based on dense blocks and GANs , and we directly insert the input image-visible light image in each layer of the entire network. We use structural similarity and gradient loss functions that are more consistent with perception instead of mean square error loss. After the adversarial training between the generator and the discriminator, we show that a trained end-to-end fusion network – the generator network – is finally obtained. Our experiments show that the fused images obtained by our approach achieve good score based on multiple evaluation indicators. Further, our fused images have better visual effects in multiple sets of contrasts, which are more satisfying to human visual perception.

[1]  Junjun Jiang,et al.  FusionGAN: A generative adversarial network for infrared and visible image fusion , 2019, Inf. Fusion.

[2]  Yu Liu,et al.  A general framework for image fusion based on multi-scale transform and sparse representation , 2015, Inf. Fusion.

[3]  Hao Zhang,et al.  Rethinking the Image Fusion: A Fast Unified Image Fusion Network based on Proportional Maintenance of Gradient and Intensity , 2020, AAAI.

[4]  Kishor P. Upla,et al.  An Edge Preserving Multiresolution Fusion: Use of Contourlet Transform and MRF Prior , 2015, IEEE Transactions on Geoscience and Remote Sensing.

[5]  Gang Liu,et al.  Multi-sensor image fusion based on fourth order partial differential equations , 2017, 2017 20th International Conference on Information Fusion (Fusion).

[6]  J. Wesley Roberts,et al.  Assessment of image fusion procedures using entropy, image quality, and multispectral classification , 2008 .

[7]  Jiayi Ma,et al.  Infrared and visible image fusion methods and applications: A survey , 2018, Inf. Fusion.

[8]  Edward H. Adelson,et al.  The Laplacian Pyramid as a Compact Image Code , 1983, IEEE Trans. Commun..

[9]  Xiao-Ping Zhang,et al.  DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion , 2020, IEEE Transactions on Image Processing.

[10]  Hui Li,et al.  Infrared and Visible Image Fusion with ResNet and zero-phase component analysis , 2018, Infrared Physics & Technology.

[11]  Paul S. Fisher,et al.  Image quality measures and their performance , 1995, IEEE Trans. Commun..

[12]  Raymond Y. K. Lau,et al.  Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[13]  Hui Li,et al.  DenseFuse: A Fusion Approach to Infrared and Visible Images , 2018, IEEE Transactions on Image Processing.

[14]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[15]  Jun Huang,et al.  Infrared and visible image fusion via saliency analysis and local edge-preserving multi-scale decomposition. , 2017, Journal of the Optical Society of America. A, Optics, image science, and vision.

[16]  Hassan Ghassemian,et al.  A new pansharpening method using multi resolution analysis framework and deep neural networks , 2017, 2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA).

[17]  David Bull,et al.  Region-Based Multimodal Image Fusion Using ICA Bases , 2007 .

[18]  Yu Liu,et al.  Multi-focus image fusion with a deep convolutional neural network , 2017, Inf. Fusion.

[19]  Laure J. Chipman,et al.  Wavelets and image fusion , 1995, Optics + Photonics.

[20]  Xiaojie Guo,et al.  U2Fusion: A Unified Unsupervised Image Fusion Network , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Wei Yu,et al.  Infrared and visible image fusion via detail preserving adversarial learning , 2020, Inf. Fusion.

[22]  Xia De-sheng Research of Measurement for Digital Image Definition , 2004 .

[23]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Haifeng Li,et al.  Dictionary learning method for joint sparse representation-based image fusion , 2013 .

[25]  Yu Han,et al.  A new image fusion performance metric based on visual information fidelity , 2013, Inf. Fusion.

[26]  Chen Chen,et al.  Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion , 2020, Inf. Fusion.

[27]  Junjun Jiang,et al.  FusionDN: A Unified Densely Connected Network for Image Fusion , 2020, AAAI.

[28]  R. Venkatesh Babu,et al.  DeepFuse: A Deep Unsupervised Approach for Exposure Fusion with Extreme Exposure Image Pairs , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[29]  Jiayi Ma,et al.  MEF-GAN: Multi-Exposure Image Fusion via Generative Adversarial Networks , 2020, IEEE Transactions on Image Processing.

[30]  Han Xu,et al.  MFF-GAN: An unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion , 2021, Inf. Fusion.

[31]  Alan C. Bovik,et al.  Image information and visual quality , 2006, IEEE Trans. Image Process..

[32]  Vladimir Petrovic,et al.  Objective image fusion performance measure , 2000 .

[33]  Hui Li,et al.  Multi-focus Image Fusion Using Dictionary Learning and Low-Rank Representation , 2017, ICIG.

[34]  Zhifeng Gao,et al.  Fusion of infrared and visible images for night-vision context enhancement. , 2016, Applied optics.

[35]  V. Aslantaş,et al.  A new image quality metric for image fusion: The sum of the correlations of differences , 2015 .

[36]  Fanjie Meng,et al.  Image fusion based on object region detection and Non-Subsampled Contourlet Transform , 2017, Comput. Electr. Eng..

[37]  Cedric Nishan Canagarajah,et al.  Pixel- and region-based image fusion with complex wavelets , 2007, Inf. Fusion.

[38]  Jan Kautz,et al.  Exposure Fusion , 2009, 15th Pacific Conference on Computer Graphics and Applications (PG'07).

[39]  Yu Zhang,et al.  Infrared and visual image fusion through infrared feature extraction and visual information preservation , 2017 .

[40]  Yang Lei,et al.  Novel fusion method for visible light and infrared images based on NSST–SF–PCNN , 2014 .

[41]  Alexei A. Efros,et al.  Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[42]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[43]  Luciano Alparone,et al.  Remote sensing image fusion using the curvelet transform , 2007, Inf. Fusion.

[44]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[45]  B. K. Shreyamsha Kumar Multifocus and multispectral image fusion based on pixel significance using discrete cosine harmonic wavelet transform , 2013 .

[46]  Alexander Toet,et al.  Image fusion by a ration of low-pass pyramid , 1989, Pattern Recognit. Lett..

[47]  Tianshuang Qiu,et al.  Medical image fusion based on sparse representation of classified image patches , 2017, Biomed. Signal Process. Control..

[48]  Toet Alexander,et al.  TNO Image Fusion Dataset , 2014 .

[49]  Wei Gao,et al.  Image fusion based on non-negative matrix factorization and infrared feature extraction , 2013, 2013 6th International Congress on Image and Signal Processing (CISP).

[50]  Ying Han,et al.  Structure-aware image fusion , 2018, Optik.

[51]  T. Durrani,et al.  NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models , 2020, IEEE Transactions on Instrumentation and Measurement.

[52]  Shutao Li,et al.  Pixel-level image fusion: A survey of the state of the art , 2017, Inf. Fusion.

[53]  Sun Li,et al.  Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters , 2016, Inf. Fusion.

[54]  Yang Chao,et al.  Efficient image fusion with approximate sparse representation , 2016, Int. J. Wavelets Multiresolution Inf. Process..

[55]  Rabab Kreidieh Ward,et al.  Deep learning for pixel-level image fusion: Recent advances and future prospects , 2018, Inf. Fusion.

[56]  Vps Naidu,et al.  Image Fusion Technique using Multi-resolution Singular Value Decomposition , 2011 .

[57]  Rick S. Blum,et al.  A categorization of multiscale-decomposition-based image fusion schemes with a performance study for a digital camera application , 1999, Proc. IEEE.

[58]  Josef Kittler,et al.  Infrared and Visible Image Fusion using a Deep Learning Framework , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[59]  Guofan Jin,et al.  Infrared and visible image fusion based on NSCT and fuzzy logic , 2010, 2010 IEEE International Conference on Mechatronics and Automation.