FuseVis: Interpreting neural networks for image fusion using per-pixel saliency visualization

Image fusion helps in merging two or more images to construct a more informative single fused image. Recently, unsupervised learning-based convolutional neural networks (CNN) have been used for different types of image-fusion tasks such as medical image fusion, infrared-visible image fusion for autonomous driving as well as multi-focus and multi-exposure image fusion for satellite imagery. However, it is challenging to analyze the reliability of these CNNs for the image-fusion tasks since no groundtruth is available. This led to the use of a wide variety of model architectures and optimization functions yielding quite different fusion results. Additionally, due to the highly opaque nature of such neural networks, it is difficult to explain the internal mechanics behind its fusion results. To overcome these challenges, we present a novel real-time visualization tool, named FuseVis, with which the end-user can compute per-pixel saliency maps that examine the influence of the input image pixels on each pixel of the fused image. We trained several image fusion-based CNNs on medical image pairs and then using our FuseVis tool we performed case studies on a specific clinical application by interpreting the saliency maps from each of the fusion methods. We specifically visualized the relative influence of each input image on the predictions of the fused image and showed that some of the evaluated image-fusion methods are better suited for the specific clinical application. To the best of our knowledge, currently, there is no approach for visual analysis of neural networks for image fusion. Therefore, this work opens a new research direction to improve the interpretability of deep fusion networks. The FuseVis tool can also be adapted in other deep neural network-based image processing applications to make them interpretable.

[1]  Simon X. Yang,et al.  A Novel approach for Multimodal Medical Image Fusion using Hybrid Fusion Algorithms for Disease Analysis , 2017 .

[2]  Junjun Jiang,et al.  FusionDN: A Unified Densely Connected Network for Image Fusion , 2020, AAAI.

[3]  Yu Liu,et al.  A general framework for image fusion based on multi-scale transform and sparse representation , 2015, Inf. Fusion.

[4]  Yu Liu,et al.  Multi-focus image fusion with a deep convolutional neural network , 2017, Inf. Fusion.

[5]  Wilfried Philips,et al.  Deep Visible and Thermal Image Fusion for Enhanced Pedestrian Visibility , 2019, Sensors.

[6]  Jan Kautz,et al.  Exposure Fusion , 2007, 15th Pacific Conference on Computer Graphics and Applications (PG'07).

[7]  Shutao Li,et al.  Multifocus image fusion by combining curvelet and wavelet transform , 2008, Pattern Recognit. Lett..

[8]  Shutao Li,et al.  The multiscale directional bilateral filter and its application to multisensor image fusion , 2012, Inf. Fusion.

[9]  Yin Fei,et al.  Medical Image Fusion Based on Feature Extraction and Sparse Representation , 2017, Int. J. Biomed. Imaging.

[10]  Cedric Nishan Canagarajah,et al.  Pixel- and region-based image fusion with complex wavelets , 2007, Inf. Fusion.

[11]  Abhishek Das,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[12]  Edmund Koch,et al.  Structural Similarity based Anatomical and Functional Brain Imaging Fusion , 2019, MBIA/MFCA@MICCAI.

[13]  Alexander Binder,et al.  On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.

[14]  B. S. Manjunath,et al.  Multisensor Image Fusion Using the Wavelet Transform , 1995, CVGIP Graph. Model. Image Process..

[15]  Bin Yang,et al.  Multi-focus image fusion and super-resolution with convolutional neural network , 2017, Int. J. Wavelets Multiresolution Inf. Process..

[16]  Ian J. Goodfellow,et al.  Efficient Per-Example Gradient Computations , 2015, ArXiv.

[17]  Junmo Kim,et al.  Generating a Fusion Image: One's Identity and Another's Shape , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[18]  Stefan Gumhold,et al.  Visualisation of Medical Image Fusion and Translation for Accurate Diagnosis of High Grade Gliomas , 2020, 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI).

[19]  Hadi Seyedarabi,et al.  A non-reference image fusion metric based on mutual information of image features , 2011, Comput. Electr. Eng..

[20]  Gemma Piella,et al.  A general framework for multiresolution image fusion: from pixels to regions , 2003, Inf. Fusion.

[21]  Xun Chen,et al.  Medical Image Fusion With Parameter-Adaptive Pulse Coupled Neural Network in Nonsubsampled Shearlet Transform Domain , 2019, IEEE Transactions on Instrumentation and Measurement.

[22]  Jinde Cao,et al.  FuseGAN: Learning to Fuse Multi-Focus Image via Conditional Generative Adversarial Network , 2019, IEEE Transactions on Multimedia.

[23]  Yi Shen,et al.  Performances evaluation of image fusion techniques based on nonlinear correlation measurement , 2004, Proceedings of the 21st IEEE Instrumentation and Measurement Technology Conference (IEEE Cat. No.04CH37510).

[24]  Yu Han,et al.  A new image fusion performance metric based on visual information fidelity , 2013, Inf. Fusion.

[25]  Xiao-Ping Zhang,et al.  DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion , 2020, IEEE Transactions on Image Processing.

[26]  Jun Wang,et al.  Image fusion with nonsubsampled contourlet transform and sparse representation , 2013, J. Electronic Imaging.

[27]  Belur V. Dasarathy,et al.  Medical Image Fusion: A survey of the state of the art , 2013, Inf. Fusion.

[28]  Lei Xiong,et al.  VIF-Net: An Unsupervised Framework for Infrared and Visible Image Fusion , 2020, IEEE Transactions on Computational Imaging.

[29]  Andrea Vedaldi,et al.  Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[30]  Henk J. A. M. Heijmans,et al.  A new quality metric for image fusion , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).

[31]  Xin Zhang,et al.  Infrared and Visible Image Fusion with a Generative Adversarial Network and a Residual Network , 2020, Applied Sciences.

[32]  Shesheng Gao,et al.  Image Segmentation-Based Multi-Focus Image Fusion Through Multi-Scale Convolutional Neural Network , 2017, IEEE Access.

[33]  Bin Yang,et al.  Image Fusion and Super-Resolution with Convolutional Neural Network , 2016, CCPR.

[34]  Ke Lu,et al.  An overview of multi-modal medical image fusion , 2016, Neurocomputing.

[35]  Rabab Kreidieh Ward,et al.  Image Fusion With Convolutional Sparse Representation , 2016, IEEE Signal Processing Letters.

[36]  Haixu Wang,et al.  Multimodal medical image fusion based on IHS and PCA , 2010 .

[37]  Zheng Liu,et al.  Directive Contrast Based Multimodal Medical Image Fusion in NSCT Domain , 2013, IEEE Transactions on Multimedia.

[38]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.

[39]  Jun Huang,et al.  A Deep Model for Multi-Focus Image Fusion Based on Gradients and Connected Regions , 2020, IEEE Access.

[40]  Rabab Kreidieh Ward,et al.  Deep learning for pixel-level image fusion: Recent advances and future prospects , 2018, Inf. Fusion.

[41]  Yi Liu,et al.  Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review , 2018, Inf. Fusion.

[42]  Q Guihong,et al.  Medical image fusion by wavelet transform modulus maxima. , 2001, Optics express.

[43]  Nick C Fox,et al.  The Alzheimer's disease neuroimaging initiative (ADNI): MRI methods , 2008, Journal of magnetic resonance imaging : JMRI.

[44]  Shutao Li,et al.  Pixel-level image fusion with simultaneous orthogonal matching pursuit , 2012, Inf. Fusion.

[45]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[46]  Junjun Jiang,et al.  FusionGAN: A generative adversarial network for infrared and visible image fusion , 2019, Inf. Fusion.

[47]  Wei Yu,et al.  Infrared and visible image fusion via detail preserving adversarial learning , 2020, Inf. Fusion.

[48]  R. Venkatesh Babu,et al.  DeepFuse: A Deep Unsupervised Approach for Exposure Fusion with Extreme Exposure Image Pairs , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[49]  B. S. Manjunath,et al.  Multi-sensor image fusion using the wavelet transform , 1994, Proceedings of 1st International Conference on Image Processing.

[50]  Bin Li,et al.  Multimodal Medical Volumetric Data Fusion Using 3-D Discrete Shearlet Transform and Global-to-Local Rule , 2014, IEEE Transactions on Biomedical Engineering.

[51]  Shutao Li,et al.  Image Fusion With Guided Filtering , 2013, IEEE Transactions on Image Processing.

[52]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[53]  Xiaojuan Ban,et al.  SESF-Fuse: An Unsupervised Deep Model for Multi-Focus Image Fusion , 2019, ArXiv.

[54]  L. Yang,et al.  Multimodality medical image fusion based on multiscale geometric analysis of contourlet transform , 2008, Neurocomputing.

[55]  Shutao Li,et al.  Pixel-level image fusion: A survey of the state of the art , 2017, Inf. Fusion.

[56]  Jingwen Yan,et al.  Image Fusion Algorithm Based on Spatia Frequency-Motivated Pulse Coupled Neural Networks in Nonsubsampled Contourlet Transform Domain: Image Fusion Algorithm Based on Spatia Frequency-Motivated Pulse Coupled Neural Networks in Nonsubsampled Contourlet Transform Domain , 2009 .

[57]  Malay Kumar Kundu,et al.  Corrections to "A Neuro-Fuzzy Approach for Medical Image Fusion" , 2015, IEEE Trans. Biomed. Eng..

[58]  Mei Yang,et al.  A novel algorithm of image fusion using shearlets , 2011 .

[59]  Vinod Kumar,et al.  Nonsubsampled shearlet based CT and MR medical image fusion using biologically inspired spiking neural network , 2015, Biomed. Signal Process. Control..

[60]  Andrew Zisserman,et al.  Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.

[61]  Shuyuan Yang,et al.  Image fusion based on a new contourlet packet , 2010, Inf. Fusion.

[62]  屈小波 Xiaobo Qu,et al.  Image Fusion Algorithm Based on Spatial Frequency-Motivated Pulse Coupled Neural Networks in Nonsubsampled Contourlet Transform Domain , 2008 .

[63]  Yu Liu,et al.  A medical image fusion method based on convolutional neural networks , 2017, 2017 20th International Conference on Information Fusion (Fusion).

[64]  Ankur Taly,et al.  Axiomatic Attribution for Deep Networks , 2017, ICML.

[65]  Avanti Shrikumar,et al.  Learning Important Features Through Propagating Activation Differences , 2017, ICML.

[66]  T. Iwatsubo [Alzheimer's disease Neuroimaging Initiative (ADNI)]. , 2011, Nihon rinsho. Japanese journal of clinical medicine.

[67]  Tong Jun-yi,et al.  フェムト秒光Kerrゲートによるイントラリピッド溶液の散乱係数の測定 | 文献情報 | J-GLOBAL 科学技術総合リンクセンター , 2011 .

[68]  Yi Chai,et al.  A novel multi-modality image fusion method based on image decomposition and sparse representation , 2017, Inf. Sci..

[69]  Syed Zulqarnain Gilani,et al.  Unsupervised Deep Multi-focus Image Fusion , 2018, ArXiv.

[70]  Shutao Li,et al.  Multimodal image fusion with joint sparsity model , 2011 .

[71]  Josef Kittler,et al.  Infrared and Visible Image Fusion using a Deep Learning Framework , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[72]  Fan Fan,et al.  A generative adversarial network with adaptive constraints for multi-focus image fusion , 2020, Neural Computing and Applications.

[73]  Yu Guo,et al.  A novel image fusion algorithm based on nonsubsampled shearlet transform , 2014 .

[74]  Bin Xiao,et al.  Union Laplacian pyramid with multiple features for medical image fusion , 2016, Neurocomputing.

[75]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[76]  LiShutao,et al.  Pixel-level image fusion , 2017 .

[77]  Hui Li,et al.  DenseFuse: A Fusion Approach to Infrared and Visible Images , 2018, IEEE Transactions on Image Processing.