A system for gaze-contingent image analysis and multi-sensorial image display

A novel system for gaze-contingent im- age analysis and multi-sensorial image display is de- scribed. The observer's scanpaths are recorded while viewing and analysing 2-0 or 3-0 (volumetric) images. A region-of-interest (ROI) centred around the current fixation point is simultaneously subjected to real-time image analysis algorithms to compute various image features, e.g. edges, textures (2-0) or surfaces and vol- umetric texture (3-0). This feature information is fed back to the observer using multiple channels, i.e. in visual (replacing the ROI by a visually modified ROI), auditory (generating an auditory display of a computed feature) and tactile (generating a tactile representation of a computed feature) manner. Thus, the observer can use several of his senses to perceive information @om the image which may be otherwise hidden to his eyes, e.g. targets or patterns which are very dificult or im- possible to detect. The human brain then fuses all the information from the multi-sensorial display. The mo- ment the eyes make a saccade to a newjixation loca- tion, the same process is applied to the new ROI centred around it. In this way the observer receives informa- tion from the local real-time image analysis around the point of gaze, hence the term gaze-contingent image analysis. The new system is profiled and several exam- ple applications are discussed.

[1]  Lester C. Loschky,et al.  Reduced saliency of peripheral targets in gaze-contingent multi-resolutional displays: blended versus sharp boundary windows , 2002, ETRA.

[2]  Carl F. R. Weiman,et al.  Video compression via log polar mapping , 1990, Defense, Security, and Sensing.

[3]  Claudio M. Privitera,et al.  Algorithms for Defining Visual Regions-of-Interest: Comparison with Eye Fixations , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[4]  M. Pomplun,et al.  Investigating the visual span in comparative search: the effects of task difficulty and divided attention , 2001, Cognition.

[5]  P Reinagel,et al.  Natural scene statistics at the centre of gaze. , 1999, Network.

[6]  David R. Bull,et al.  Gaze-Contingent Multi-modality Displays of Multi-layered Geographical Maps , 2002, Numerical Methods and Application.

[7]  G. McConkie,et al.  The span of the effective stimulus during a fixation in reading , 1975 .

[8]  R. Carpenter,et al.  Movements of the Eyes , 1978 .

[9]  S. G. Nikolov,et al.  Volume visualisation via region enhancement around an observer's fixation point , 2000 .

[10]  Stavri G. Nikolov,et al.  Multi-modality gaze-contingent displays for image fusion , 2002, Proceedings of the Fifth International Conference on Information Fusion. FUSION 2002. (IEEE Cat.No.02EX5997).

[11]  James M. Hillis,et al.  Combining Sensory Information: Mandatory Fusion Within, but Not Between, Senses , 2002, Science.

[12]  Lester C. Loschky,et al.  User Performance With Gaze Contingent Displays , 2022 .

[13]  Lester C. Loschky,et al.  User performance with gaze contingent multiresolutional displays , 2000, ETRA.

[14]  K. Rayner Eye movements in reading and information processing: 20 years of research. , 1998, Psychological bulletin.

[15]  Trevor Darrell,et al.  Signal level fusion for multimodal perceptual user interface , 2001, PUI '01.

[16]  M. Ernst,et al.  Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.

[17]  Peter De Graef,et al.  On-line control of moving masks and windows on a complex background using the ATVista videographics adapter , 1994 .

[18]  Mark Segal,et al.  The OpenGL Graphics System: A Specification , 2004 .

[19]  Wilson S. Geisler,et al.  Implementation of a foveated image coding system for image bandwidth reduction , 1996, Electronic Imaging.

[20]  S Saida,et al.  Useful visual field size for pattern perception , 1979, Perception & psychophysics.