A retina based multi-resolution image-fusion

Multiresolution images can be merged more efficiently, by taking into account the multiscale properties of the human vision system. To achieve this goal it is necessary to build a multiscale fusion model, based on the retinal cell photoreceptors. This paper introduces a general issue of this model, and the new application of the model in multispectral image fusion. The proposed method is compared with the HSI, PCA, Brovey, and wavelet, methods. Results show it preserves more spectral features with less spatial distortion.