Accurate mapping of natural scenes radiance to cone activation space: a new image dataset

The characterization of trichromatic cameras is usually done in terms of a device-independent color space, such as the CIE 1931 XYZ space. This is indeed convenient since it allows the testing of results against colorimetric measures. We have characterized our camera to represent human cone activation by mapping the camera sensor’s (RGB) responses to human (LMS) through a polynomial transformation, which can be “customized” according to the types of scenes we want to represent. Here we present a method to test the accuracy of the camera measures and a study on how the choice of training reflectances for the polynomial may alter the results. Introduction The last decade has seen an increasing interest in the interplay between the distinctive characteristics of biological sensory systems and those of the natural environment. In the case of vision, this interest reflects the growing evidence that the statistical properties (both spatial and chromatic) of the visual environment have contributed to shape the way in which our visual system (and that of other species) function. Consequently, much research is based on the analysis of the visual environment (considering the tasks that a living organism needs to perform in order to survive and its biological constraints) with the aim of learning about the statistical regularities that the visual system may have exploited in its development. In his review work on the relationships between visual perception and the statistical properties of natural scenes, Geisler 1 points out that measuring within-domain statistics is central to testing for “efficient coding” (the hypothesis that the response characteristics of visual neurons can be predicted from the statistics of natural images plus some biological constraints). Only after knowing the probability distribution of the property considered, we can determine which is the most efficient way of coding it. To this respect it is important to point out that there is an exponential relationship between the number of samples required to estimate a probability distribution and the number of properties considered. In other words, the more complex regularities of the visual environment we want to map, the larger the number of scenes we need to gather: this is the main reason why scientists have so far concentrated on only a small group of properties which need few images to compute. Other reason has to do with technical limitations: when natural scene regularities are unrelated to the chromatic responses of the visual system (multiscale analysis, contours, etc.), the use of uncalibrated imagery is justified. However when they involve the chromatic domain, a more sophisticated approach is needed. Two techniques and methods were initially tried to measure and compute the statistical regularities of nature in the chromatic domain: (a) spectroradiometric devices which measure spectral radiance (radiance as a function of wavelength) from a small patch of image at the time, obtaining information about illuminants and reflective material properties; (b) hyperspectral cameras which measure the same from a whole image at the time but require long exposures, etc. Both these methods are impractical for gathering large databases of in-the-field imagery: spectroradiometric devices do not capture the spatial properties of natural images and hyperspectral cameras are only useful for indoor environments or when there is little change in time (long distance shots, man-made structures, landscapes, etc.). A third method has been tried more recently to reach a compromise between speed, portability and accuracy: calibrated trichromatic cameras are fast and portable but do not provide the complete spectral information necessary to fully characterize the reflectance of every patch of the image, however they are the only way to record the statistics of large samples of the visual environment to date. The latest advances in digital imaging have turned trichromatic cameras into the most common device for estimating/measuring the properties of natural scenes. Commercial digital cameras are relatively cheap and if properly calibrated they can provide photometric information for every region of the scene (i.e. a measure of the radiant power absorbed by each of the camera’s sensors, for each pixel). Calibrating a digital camera generally involves converting the image captured in the camera (also called device-dependent) color space into a reference (or device-independent) color-space. There are currently several methods to produce this color space transformation (see Martinez-Verdú et al for a more detailed explanation). One of such methods (the spectroradiometric approach) consists of obtaining the camera sensor’s response to a narrowband monochromatic stimulus which is in turn varied to span the whole spectral sensitivity range of the camera. Since the stimulus’ radiometric characteristics are known, it is possible to reconstruct the sensor’s spectral sensitivity at each wavelength (and for each color sensor). It is also necessary to measure the sensor’s output dependency on radiant power to obtain a complete picture of the camera’s response to light. Once the sensor’s responses are known, it is possible to find an approximate way of transforming the camera’s RGB values to any device-independent color space (commonly the CIE 1931 XYZ color space). A second approach to the characterization of digital cameras is based on mathematical models (mathematical approach) which estimate the camera’s matching functions from the device RGB responses to a set of (known) spectral reflectances, such as the squares of the Macbeth ColorChecker card. While the first approach is quite precise, it is seldom used because of its complexity. The last method is easier to implement but it is very vulnerable to measurement noise. Some intermediate approaches rely on assigning an estimated function to the camera’s sensors and performing a mapping of the camera’s space by means of a “training set” of RGB responses and radiometric measurements. Our approach is a mixture of the two: it consists of measuring the camera sensitivities by means of photographing a white target through a set of spectrally narrowband interference filters (spectroradiometric approach) while using a training set to “match” the theoretical camera output to a device independent space (mathematical approach). 50 ©2010 Society for Imaging Science and Technology The problem of converting to cone activation space Whatever solution is chosen for characterizing the camera output in terms of a device-independent color space, the further transformation of these values into a cone activation space is not without difficulties. Cone activation spaces are physiologically realistic alternatives to the already ubiquitous systems of specifying color adopted by the Commission Internationale de l’Eclairage (CIE), being the best established of these systems the CIE 1931. When the CIE 1931 systems was adopted, the spectral sensitivities of the actual photoreceptors in the human retina were not indisputably known and instead, a set of hypothetical primaries was adopted, based on the experiments of Guild and Wright to determine the human color matching functions. The trichromatic values XYZ of the CIE 1931 system can be understood as the photon catches of three arbitrary photoreceptors with spectral sensitivities determined by the so called , , x y z functions. These functions are approximately point-by-point linear transformations of the cone spectral sensitivities of an average human observer (in fact z is actually very close to the spectral sensitivity of human short-wavelength -or “S” cones and y was chosen to have the same shape as the standard function of luminous sensitivity or V ). Despite the CIE 1931’s popularity and some obvious advantages, a chromaticity system that is not based on human physiology (or any other physiology, as in this case) is of limited use for researching the neural properties of a visual system. To amend this situation, a number of physiologically-plausible chromatic systems have been adopted by the neuroscience community, being one of the most popular the MacLeodBoynton space. In the MacLeod-Boynton space, the axes correspond to two of the chromatic channels identified physiologically by Derrington et al 17 in the early visual system. In this space, physiologically significant loci are represented by horizontal and vertical lines. To make the situation more complicated, MacLeod-Boynton system is derived from the Smith and Pokorny human cone sensitivities, which in turn are not exact point-by-point transformations of the CIE , , x y z , but of the slightly different set of primaries calculated by Judd in 1951 (and tabulated by Vos in 1978) known as the Judd-Vos response functions. These are favored in visual science because of its better estimate of luminosity at short wavelengths. There is a formula for transforming between the chromaticity coordinates of the CIE 1931 and the Judd 1951 system but it is valid only for monochromatic lights. This means that to use the MacLeod-Boynton system or any other cone activation space derived from the Smith and Pokorny (1975) sensitivities it is necessary to know the spectroradiometric properties of the stimulus. The most straightforward way of avoiding the inconvenience of a two-part chromatic conversion of the stimulus (from device-dependent camera-RGB space to CIE 1931 XYZ and then to Smith and Pokorny LMS (L for long, M for middle and S for short wavelength) cone activation space with the consequent transformation errors, is to characterize the camera directly in terms of the later (LMS) space. This can be done if one already knows the camera’s sensor spectral sensitivities by means of finding the best transformation between the two chromatic systems. In this work we have based our analysis in the Smith and Pokorny (1975) cone responses, which are calculated at t

[1]  P. Capilla,et al.  Calculation of the color matching functions of digital cameras from their complete spectral sensitivities : Color imaging science , 2002 .

[2]  W K Pratt,et al.  Spectral estimation techniques for the spectral calibration of a color image scanner. , 1976, Applied optics.

[3]  Peter A. Rhodes,et al.  A study of digital camera colorimetric characterisation based on polynomial modelling , 2001 .

[4]  A. Stockman,et al.  The spectral sensitivities of the middle- and long-wavelength-sensitive cones derived from measurements in observers of known genotype , 2000, Vision Research.

[5]  J. Pokorny,et al.  Spectral sensitivity of the foveal cone photopigments between 400 and 500 nm , 1975, Vision Research.

[6]  H. Trussell Applications of set theoretic methods to color systems , 1991 .

[7]  Paul M. Hubel,et al.  Eyeing the Camera: into the Next Century , 2002, Color Imaging Conference.

[8]  H. Joel Trussell,et al.  Mathematical methods for the analysis of color scanning filters , 1997, IEEE Trans. Image Process..

[9]  Flávio P. Ferreira,et al.  Statistics of spatial cone-excitation ratios in natural scenes. , 2002, Journal of the Optical Society of America. A, Optics, image science, and vision.

[10]  P. Lennie,et al.  Chromatic mechanisms in lateral geniculate nucleus of macaque. , 1984, The Journal of physiology.

[11]  Peter Morovic,et al.  Metamer-set-based approach to estimating surface reflectance from camera RGB. , 2006, Journal of the Optical Society of America. A, Optics, image science, and vision.

[12]  H. Ives The transformation of color-mixture equations from one system to another , 1915 .

[13]  Tony Johnson,et al.  Methods for characterizing colour scanners and digital cameras , 1996 .

[14]  Yung-Chang Chen,et al.  Colorimetric modeling for vision systems , 2000, J. Electronic Imaging.

[15]  M. Kuittinen,et al.  Spectral representation of color images , 1988, [1988 Proceedings] 9th International Conference on Pattern Recognition.

[16]  Stephen Westland,et al.  A comparative study of the characterisation of colour cameras by means of neural networks and polynomial transforms , 2004 .

[17]  Roy S. Berns,et al.  Colorimetric characterization of a desktop drum scanner using a spectral model , 1995, J. Electronic Imaging.

[18]  W. Geisler Visual perception and the statistical properties of natural scenes. , 2008, Annual review of psychology.

[19]  Joseph Ernest Petavel,et al.  The colorimetric properties of the spectrum , 1931 .

[20]  R. M. Boynton,et al.  Chromaticity diagram showing cone excitation by stimuli of equal luminance. , 1979, Journal of the Optical Society of America.

[21]  Francisco Martínez-Verdú,et al.  Characterization of a digital camera as an absolute tristimulus colorimeter , 2003, IS&T/SPIE Electronic Imaging.

[22]  Kinjiro Amano,et al.  Information limits on neural identification of colored surfaces in natural scenes , 2004, Visual Neuroscience.

[23]  Po-chieh Hung,et al.  Colorimetric calibration for scanners and media , 1991, Electronic Imaging.

[24]  Shoji Tominaga,et al.  MULTICHANNEL VISION SYSTEM FOR ESTIMATING SURFACE AND ILLUMINATION FUNCTIONS , 1996 .

[25]  H. Joel Trussell,et al.  Color device calibration: a mathematical formulation , 1999, IEEE Trans. Image Process..

[26]  J. J. Vos Colorimetric and photometric properties of a 2° fundamental observer , 1978 .

[27]  L. Maloney,et al.  Color constancy: a method for recovering surface spectral reflectance. , 1986, Journal of the Optical Society of America. A, Optics and image science.

[28]  Po-Chieh Hung,et al.  Colorimetric calibration in electronic imaging devices using a look-up-table model and interpolations , 1993, J. Electronic Imaging.

[29]  W. D. Wright A re-determination of the trichromatic coefficients of the spectral colours , 1929 .

[30]  Mark S. Drew,et al.  Constrained least-squares regression in color spaces , 1997, J. Electronic Imaging.