Semantic Image Fusion

Image fusion methods and metrics for their evaluation have conventionally used pixel based or low level features. However, for many applications the aim of image fusion is to effectively combine the semantic content of the input images. This paper proposes a novel system for the semantic combination of visual content using pre-trained CNN network architectures. Our proposed semantic fusion is initiated through the fusion of the top layer feature map outputs (for each input image) through gradient updating of the fused image input (so called image optimisation). Simple ‘choose maximum’ and ‘local majority’ filter based fusion rules are utilised for feature map fusion. This provides a simple method to combine layer outputs and thus a unique framework to fuse single channel and colour images within a decomposition pre-trained for classification and therefore aligned with semantic fusion. Furthermore, class activation mappings of each input image are used to combine semantic information at a higher level. The developed methods are able to give equivalent low level fusion performance to state of the art methods while providing a unique architecture to combine semantic information from multiple images.

[1]  Zhenfeng Shao,et al.  Remote Sensing Image Fusion With Deep Convolutional Neural Network , 2018, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[2]  Yun He,et al.  A multiscale approach to pixel-level image fusion , 2005, Integr. Comput. Aided Eng..

[3]  Guoyin Wang,et al.  Pixel convolutional neural network for multi-focus image fusion , 2017, Inf. Sci..

[4]  Vladimir Petrovic,et al.  Objective image fusion performance measure , 2000 .

[5]  拓海 杉山,et al.  “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”の学習報告 , 2017 .

[6]  Symeon Chatzinotas,et al.  IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing , 2020 .

[7]  Hassan Ghassemian,et al.  A review of remote sensing image fusion methods , 2016, Inf. Fusion.

[8]  Wassim Hamidouche,et al.  Reveal of Vision Transformers Robustness against Adversarial Attacks , 2021, ArXiv.

[9]  Ali Aghagolzadeh,et al.  Ensemble of CNN for multi-focus image fusion , 2019, Inf. Fusion.

[10]  Eero P. Simoncelli,et al.  A Parametric Texture Model Based on Joint Statistics of Complex Wavelet Coefficients , 2000, International Journal of Computer Vision.

[11]  Cedric Nishan Canagarajah,et al.  Image Fusion Using Complex Wavelets , 2002, BMVC.

[12]  Changsheng Xu,et al.  StyTr^2: Unbiased Image Style Transfer with Transformers , 2021, ArXiv.

[13]  Zhifei Zhang,et al.  A Semantic-based Medical Image Fusion Approach , 2019, ArXiv.

[14]  Farhad Samadzadegan,et al.  A review of image fusion techniques for pan-sharpening of high-resolution satellite imagery , 2021 .

[15]  Hang Zhang,et al.  Multi-style Generative Network for Real-time Transfer , 2017, ECCV Workshops.

[16]  Shutao Li,et al.  The multiscale directional bilateral filter and its application to multisensor image fusion , 2012, Inf. Fusion.

[17]  Josef Kittler,et al.  Infrared and Visible Image Fusion using a Deep Learning Framework , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[18]  Zunlei Feng,et al.  Neural Style Transfer: A Review , 2017, IEEE Transactions on Visualization and Computer Graphics.

[19]  Kishore Rajendiran,et al.  Multi sensor image fusion for surveillance applications using hybrid image fusion algorithm , 2017, Multimedia Tools and Applications.

[20]  Xin Lin,et al.  Style Transfer for Anime Sketches with Enhanced Residual U-net and Auxiliary Classifier GAN , 2017, 2017 4th IAPR Asian Conference on Pattern Recognition (ACPR).

[21]  Shuyuan Yang,et al.  Image fusion based on a new contourlet packet , 2010, Inf. Fusion.

[22]  T. Durrani,et al.  NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models , 2020, IEEE Transactions on Instrumentation and Measurement.

[23]  Antonio J. Plaza,et al.  Remote Sensing Image Fusion Using Hierarchical Multimodal Probabilistic Latent Semantic Analysis , 2018, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing.

[24]  Belur V. Dasarathy,et al.  Medical Image Fusion: A survey of the state of the art , 2013, Inf. Fusion.

[25]  Leon A. Gatys,et al.  Texture synthesis and the controlled generation of natural stimuli using convolutional neural networks , 2015, ArXiv.

[26]  Leon A. Gatys,et al.  A Neural Algorithm of Artistic Style , 2015, ArXiv.

[27]  Xiaojie Guo,et al.  U2Fusion: A Unified Unsupervised Image Fusion Network , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[28]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[29]  Abhishek Das,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[30]  Shutao Li,et al.  Pixel-level image fusion: A survey of the state of the art , 2017, Inf. Fusion.

[31]  Bolei Zhou,et al.  Learning Deep Features for Discriminative Localization , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  Andrea Vedaldi,et al.  Texture Networks: Feed-forward Synthesis of Textures and Stylized Images , 2016, ICML.

[33]  Chuan Li,et al.  Combining Markov Random Fields and Convolutional Neural Networks for Image Synthesis , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Jong Beom Ra,et al.  Contrast-Enhanced Fusion of Multisensor Images Using Subband-Decomposed Multiscale Retinex , 2012, IEEE Transactions on Image Processing.

[35]  Mohammed Yeasin,et al.  Eigen-CAM: Class Activation Map using Principal Components , 2020, 2020 International Joint Conference on Neural Networks (IJCNN).

[36]  Li Fei-Fei,et al.  Perceptual Losses for Real-Time Style Transfer and Super-Resolution , 2016, ECCV.

[37]  Ming-Hsuan Yang,et al.  Diversified Texture Synthesis with Feed-Forward Networks , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[38]  Ezzeddine Zagrouba,et al.  Multimodal medical image fusion review: Theoretical background and recent advances , 2021, Signal Process..

[39]  Nantheera Anantrasirichai,et al.  Image Fusion via Sparse Regularization with Non-Convex Penalties , 2020, Pattern Recognit. Lett..

[40]  Jonathon Shlens,et al.  A Learned Representation For Artistic Style , 2016, ICLR.

[41]  Henk J. A. M. Heijmans,et al.  A new quality metric for image fusion , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).