Towards cognitive image fusion

The increasing availability and deployment of imaging sensors operating in multiple spectral bands has led to a large research effort in image fusion, resulting in a plethora of pixel-level image fusion algorithms. However, the cognitive aspects of multisensor image fusion have not received much attention in the development of these methods. In this study we investigate how humans interpret visual and infrared images, and we compare the interpretation of these individual image modalities to their fused counterparts, for different image fusion schemes. This was done in an attempt to test to what degree image fusion schemes can enhance human perception of the structural layout and composition of realistic outdoor scenes. We asked human observers to manually segment the details they perceived as most prominent in a set of corresponding visual, infrared and fused images. For each scene, the segmentations of the individual input image modalities were used to derive a joint reference (“gold standard”) contour image that represents the visually most salient details from both of these modalities and for that particular scene. The resulting reference images were then used to evaluate the manual segmentations of the fused images, using a precision-recall measure as the evaluation criterion. In this sense, the best fusion method provides the largest number of correctly perceived details (originating from each of the individual modalities that were used as input for the fusion scheme) and the smallest amount of false alarms (fusion artifacts or illusory details). The results show that this evaluation method can provide valuable insight into the way fusion schemes combine perceptually important details from the individual input image modalities. Given a reference contour image, the method can potentially be used to design image fusion schemes that are optimally tuned to human visual perception for different applications and scenarios (e.g. environmental or weather conditions).

[1]  Jan Noyes,et al.  Psychophysical and metric assessment of fused images , 2005, APGV '05.

[2]  Moira I. Smith,et al.  Image fusion technology for security and surveillance applications , 2006, SPIE Security + Defence.

[3]  Roberto de Alencar Lotufo,et al.  Fast multidimensional parallel Euclidean distance transform based on mathematical morphology , 2001, Proceedings XIV Brazilian Symposium on Computer Graphics and Image Processing.

[4]  Pramod K. Varshney,et al.  A human perception inspired quality metric for image fusion based on regional information , 2007, Inf. Fusion.

[5]  Jan Noyes,et al.  Scanpath Analysis of Fused Multi-Sensor Images with Luminance Change: A Pilot Study , 2006, 2006 9th International Conference on Information Fusion.

[6]  Xin Liu,et al.  A novel similarity based quality metric for image fusion , 2008, Inf. Fusion.

[7]  James W. Davis,et al.  Background-subtraction using contour-based fusion of thermal and visible imagery , 2007, Comput. Vis. Image Underst..

[8]  Alexander Toet,et al.  Uni-Modal Versus Joint Segmentation for Region-Based Image Fusion , 2006, 2006 9th International Conference on Information Fusion.

[9]  Yonghong Jia,et al.  A method based on IHS cylindrical transform model for quality assessment of image fusion , 2005, International Symposium on Multispectral Image Processing and Pattern Recognition.

[10]  Stavri G. Nikolov,et al.  Assessment of images fused using false colouring , 2010 .

[11]  Fernando Pereira,et al.  Methodologies for objective evaluation of video segmentation quality , 2003, Visual Communications and Image Processing.

[12]  Cedric Nishan Canagarajah,et al.  A Novel Metric for Performance Evaluation of Image Fusion Algorithms , 2005, IEC.

[13]  Marco Diani,et al.  Enhancement of Sight Effectiveness by Dual Infrared System: Evaluation of Image Fusion Strategies , 2005 .

[14]  Nasser Kehtarnavaz,et al.  Real-Time Imaging VI , 2002 .

[15]  Belur V. Dasarathy,et al.  Sensor Fusion: Architectures, Algorithms, and Applications , 1998 .

[16]  Bernt Schiele,et al.  International Journal of Computer Vision manuscript No. (will be inserted by the editor) Semantic Modeling of Natural Scenes for Content-Based Image Retrieval , 2022 .

[17]  Jason S. McCarley,et al.  Perceptual Ability with Real-World Nighttime Scenes: Image-Intensified, Infrared, and Fused-Color Imagery , 1999, Hum. Factors.

[18]  Jean Serra,et al.  Image Analysis and Mathematical Morphology , 1983 .

[19]  Edward A Essock,et al.  Human perceptual performance with nonliteral imagery: region recognition and texture-based segmentation. , 2004, Journal of experimental psychology. Applied.

[20]  Petros Maragos,et al.  Morphological skeleton representation and coding of binary images , 1984, IEEE Trans. Acoust. Speech Signal Process..

[21]  Jitendra Malik,et al.  Learning to detect natural image boundaries using local brightness, color, and texture cues , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  Cedric Nishan Canagarajah,et al.  Pixel- and region-based image fusion with complex wavelets , 2007, Inf. Fusion.

[23]  Yi Shen,et al.  Performance Assessment of Image Fusion , 2006, PSIVT.

[24]  Fernando Pereira,et al.  Video Object Relevance Metrics for Overall Segmentation Quality Evaluation , 2006, EURASIP J. Adv. Signal Process..

[25]  Henk J. A. M. Heijmans,et al.  A new quality metric for image fusion , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).

[26]  Rick S. Blum,et al.  Multi-sensor image fusion and its applications , 2005 .

[27]  Yufeng Zheng,et al.  A new metric based on extended spatial frequency and its application to DWT based fusion algorithms , 2007, Inf. Fusion.

[28]  Bernd Korn Enhanced and Synthetic Vision , 2006 .

[29]  Jamie P. Heather,et al.  A review of image fusion technology in 2005 , 2005, SPIE Defense + Commercial Sensing.

[30]  John M. Irvine,et al.  Information Fusion for Feature Extraction and the Development of Geospatial Information , 2004 .

[31]  Allen M. Waxman,et al.  Field evaluations of dual-band fusion for color night vision , 1999, Defense, Security, and Sensing.

[32]  Bir Bhanu,et al.  Tracking Humans using Multi-modal Fusion , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[33]  Gerald C. Holst,et al.  Infrared Imaging Systems: Design, Analysis, Modeling, and Testing XXX , 1990 .

[34]  Luis Rueda,et al.  Advances in Image and Video Technology, Second Pacific Rim Symposium, PSIVT 2007, Santiago, Chile, December 17-19, 2007, Proceedings , 2007, PSIVT.

[35]  Carl E. Halford,et al.  LWIR and MWIR fusion algorithm comparison using image metrics , 2005, SPIE Defense + Commercial Sensing.

[36]  Seong G. Kong,et al.  Multiscale Fusion of Visible and Thermal IR Images for Illumination-Invariant Face Recognition , 2007, International Journal of Computer Vision.

[37]  Allen M. Waxman,et al.  Multisensor image fusion and mining: learning targets across extended operating conditions , 2004, SPIE Defense + Commercial Sensing.

[38]  Alexander Toet,et al.  Perceptual evaluation of different image fusion schemes , 2003 .

[39]  Fernando Pereira,et al.  Stand-Alone Objective Segmentation Quality Evaluation , 2002, EURASIP J. Adv. Signal Process..

[40]  Pramod K. Varshney,et al.  A perceptual quality metric for image fusion based on regional information , 2005, SPIE Defense + Commercial Sensing.

[41]  Vladimir S. Petrovic,et al.  Objective pixel-level image fusion performance measure , 2000, SPIE Defense + Commercial Sensing.

[42]  Colin M. Lewis Optics and photonics for counterterrorism and crime fighting III : 18-20 September 2007, Florence, Italy , 2006 .

[43]  Ivan Kadar,et al.  Signal Processing, Sensor Fusion, and Target Recognition XV , 1994 .

[44]  Nishan Canagarajah,et al.  A Similarity Metric for Assessment of Image Fusion Algorithms , 2008 .

[45]  Rick S. Blum On multisensor image fusion performance limits from an estimation theory perspective , 2006, Inf. Fusion.

[46]  Jacques Verly Enhanced and Synthetic Vision 1997 , 1996 .

[47]  M. E. Ulug,et al.  Quantitative metric for comparison of night vision fusion algorithms , 2000, SPIE Defense + Commercial Sensing.

[48]  William K. Krebs,et al.  Psychophysical Assessments of Image-Sensor Fused Imagery , 2002, Hum. Factors.

[49]  Vassilis Tsagaris,et al.  Information measure for assessing pixel-level fusion methods , 2004, SPIE Remote Sensing.

[50]  Jitendra Malik,et al.  A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics , 2001, Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001.

[51]  C. Angell Fusion performance using a validation approach , 2005, 2005 7th International Conference on Information Fusion.

[52]  D. A. Fay,et al.  Multisensor image fusion & mining: from neural systems to COTS software , 2003, IEMC '03 Proceedings. Managing Technologically Driven Organizations: The Human Side of Innovation and Change (IEEE Cat. No.03CH37502).

[53]  Alexander Toet,et al.  Performance comparison of different gray-level image fusion schemes through a universal image quality index , 2003, SPIE Defense + Commercial Sensing.

[54]  Allen M. Waxman,et al.  Sensor Fused Night Vision : Assessing Image Quality in the Lab and in the Field , 2006, 2006 9th International Conference on Information Fusion.

[55]  Alexander Toet,et al.  Fusion of visible and thermal imagery improves situational awareness , 1997, Defense, Security, and Sensing.

[56]  S. Narayanan,et al.  Cognitively-engineered multisensor image fusion for military applications , 2009, Inf. Fusion.

[57]  S. Ullman Object recognition and segmentation by a fragment-based hierarchy , 2007, Trends in Cognitive Sciences.

[58]  Sebastiano B. Serpico,et al.  Image and Signal Processing for Remote Sensing V , 1998 .

[59]  David Hooper,et al.  Real-time image fusion: a vision aid for helicopter pilotage , 2002, IS&T/SPIE Electronic Imaging.