How the Geometry of Space controls Visual Attention during Spatial Decision Making Jan M. Wiener (jan.wiener@cognition.uni-freiburg.de) Christoph H¨olscher (christoph.hoelscher@cognition.uni-freiburg.de) Simon B uchner (simon.buechner@cognition.uni-freiburg.de) Lars Konieczny (lars.konieczny@cognition.uni-freiburg.de) Center for Cognitive Science, Freiburg University, Friedrichstr. 50, D-79098 Freiburg, Germany Abstract In this paper we present an eye-tracking experiment investi- gating the control of visual attention during spatial decision making. Participants were presented with screenshots taken at different choice points in a large complex virtual indoor en- vironment. Each screenshot depicted two movement options. Participants had to decide between them in order to search for an object that was hidden in the environment. We demonstrate (1.) that participants reliably chose the movement option that featured the longest line of sight, (2.) a robust gaze bias to- wards the eventually chosen movement option, and (3.) using a bottom-up description that captures aspects of the geometry of the sceneries depicted, we were able to predict participants’ fixation behavior. Taken together, results from this study shed light onto the control of visual attention during navigation and wayfinding. Keywords: visual attention; wayfinding; navigation; gaze be- havior; spatial cognition; spatial perception. Introduction What controls visual attention when navigating through space? In the context of navigation, eye-tracking studies so far primarily investigated the role of gaze for the control of locomotory or steering behavior (Grasso, Prevost, Ivanenko, & Berthoz, 1998; Hollands, Patla, & Vickers, 2002; Wilkie & Wann, 2003). Wayfinding, however, also includes pro- cesses such as encoding and retrieving information from spa- tial memory, path planning, and spatial decision making at choice points (c.f. Montello, 2001). So far, very few, if any, studies made use of eye-tracking techniques to investi- gate such higher level cognitive processes involved in navi- gation and wayfinding. For example, which information do navigators attend to and process when deciding between path alternatives? And, how does gaze behavior relate to spatial decision making at all? To approach these questions we pre- sented participants with images of choice points and asked them to decide between two movement options while record- ing their eye-movements. In non-spatial contexts, gaze behavior has been shown to reflect preferences in visual decision tasks (Glaholt & Rein- gold, in press). In two alternative forced choice paradigms in which participants have to judge attractiveness of faces, for example, gaze probability is initially distributed equally between alternatives. Only briefly before the decision, gaze gradually shifts towards the eventually chosen stimulus (Shi- mojo, Simion, Shimojo, & Scheier, 2003; Simion & Shimojo, 2007). It is an open question whether similar effects can also be observed in spatial decision making such as path choice behavior. The features people attend to when inspecting images of scenes have been investigated in numerous studies revealing both, bottom-up (stimulus derived) as well as of top-down (e.g., task) influences (for an overview see Henderson, 2003). Already in the 60s, Yarbus (1967) demonstrated influences of the task on the control of visual attention: participants’ gaze patterns when inspecting the same drawing systemati- cally differed when asked to judge the ages of people depicted or when asked to estimate their material circumstances. The most widely used bottom-up approach is that of saliency maps (Itti & Koch, 2000, 2001). A saliency map is a representa- tion of the stimulus in which the strength of different fea- tures (color, intensity, orientation) are coded. Several studies demonstrated that saliency maps are useful predictors of early fixations, particularly when viewing natural complex scenes (e.g., Foulsham & Underwood, 2008). It is important to stress that bottom-up approaches usu- ally do not explicitly account for the fact that images or pic- tures are two-dimensional projections of three-dimensional scenes. In other words, the geometrical properties of the scenes depicted in the images are not necessarily captured or highlighted by, for example, saliency maps. For naviga- tion and wayfinding, however, the interpretation and under- standing of the depicted three dimensional structure may be inevitable. This opens up intriguing questions: Is it possible to predict gaze behavior by analyzing geometrical properties of the sceneries depicted if the viewer is solving a navigation task? If so, can the analysis of gaze behavior be used to infer the strategies and heuristics underlying different navigation or wayfinding tasks? And, which kind of description systems of spatial form and structure captures properties of space that are relevant for the control of visual attention? Promising candidates are isovists or viewshed polygons (Benedikt, 1979), which both describe the visible area from the perspective of the observer. Isovists are essentially depth profiles and several quantitative descriptors such as the visi- ble area, the length of the perimeter, the number of vertices, etc., can be derived that reflect local physical properties of the corresponding space. Moreover, isovists have been shown to capture properties of the geometry of environments that are relevant for experience of the corresponding space and locomotion within the space (Wiener et al., 2007; Franz & Wiener, 2008). The specific research questions for this study were as fol- lows:
[1]
C. Koch,et al.
A saliency-based search mechanism for overt and covert shifts of visual attention
,
2000,
Vision Research.
[2]
A. L. Yarbus,et al.
Eye Movements and Vision
,
1967,
Springer US.
[3]
Antonio Torralba,et al.
Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search.
,
2006,
Psychological review.
[4]
M. Benedikt,et al.
To Take Hold of Space: Isovists and Isovist Fields
,
1979
.
[5]
A. Berthoz,et al.
Eye-head coordination for the steering of locomotion in humans: an anticipatory synergy
,
1998,
Neuroscience Letters.
[6]
C. Koch,et al.
Computational modelling of visual attention
,
2001,
Nature Reviews Neuroscience.
[7]
Wilfried Brauer,et al.
Spatial Cognition III
,
2003,
Lecture Notes in Computer Science.
[8]
T. Foulsham,et al.
What can saliency models predict about eye movements? Spatial and sequential aspects of fixations during encoding and recognition.
,
2008,
Journal of vision.
[9]
Yoshiaki Shirai,et al.
Look where you're going [robotic wheelchair]
,
2003,
IEEE Robotics Autom. Mag..
[10]
Eyal M. Reingold,et al.
The time course of gaze bias in visual decision tasks
,
2009
.
[11]
A. Patla,et al.
“Look where you’re going!”: gaze behaviour associated with maintaining and changing the direction of locomotion
,
2002,
Experimental Brain Research.
[12]
S. Shimojo,et al.
Gaze bias both reflects and influences preference
,
2003,
Nature Neuroscience.
[13]
Jan M. Wiener,et al.
From Space Syntax to Space Semantics: A Behaviorally and Perceptually Oriented Methodology for the Efficient Description of the Geometry and Topology of Environments
,
2008
.
[14]
Zijiang J. He,et al.
Distance determined by the angular declination below the horizon
,
2001,
Nature.
[15]
J. Henderson.
Human gaze control during real-world scene perception
,
2003,
Trends in Cognitive Sciences.
[16]
R Conroy-Dalton.
The secret is to follow your nose: route path selection and angularity
,
2001
.
[17]
Richard M Wilkie,et al.
Eye-movements aid the control of locomotion.
,
2003,
Journal of vision.
[18]
Shinsuke Shimojo,et al.
Interrupting the cascade: Orienting contributes to decision making even in the absence of visual stimulation
,
2007,
Perception & psychophysics.
[19]
A. L. I︠A︡rbus.
Eye Movements and Vision
,
1967
.
[20]
Yong Jeong,et al.
Spatial cognition
,
2004,
Neurology.
[21]
Heinrich H Bülthoff,et al.
Isovist Analysis Captures Properties of Space Relevant for Locomotion and Experience
,
2007,
Perception.