A computational theory of human perceptual mapping

A computational theory of human perceptual mapping W. K. Yeap (wai.yeap@aut.ac.nz) Centre for Artificial Intelligence Research Auckland University of Technology, New Zealand Abstract However, these views need to be organized into a coherent global map and a method has been suggested. It requires recognizing objects found in the selected views in all the in- between views that have not been selected. These objects allow one to triangulate one’s position in the map and add new views to the map in their appropriate position. The theory has been tested successfully with different implementations on mobile robots and the resulting maps produced were found to exhibit several interesting characteristics of a human perceptual map. This paper presents a new computational theory of how humans integrate successive views to form a perceptual map. Traditionally, this problem has been thought of as a straightforward integration problem whereby position of objects in one view is transformed to the next and combined. However, this step creates a paradoxical situation in human perceptual mapping. On the one hand, the method requires errors to be corrected and the map to be constantly updated, and yet, on the other hand, human perception and memory show a high tolerance for errors and little integration of successive views. A new theory is presented which argues that our perceptual map is computed by combining views only at their limiting points. To do so, one must be able to recognize and track familiar objects across views. The theory has been tested successfully on mobile robots and the lessons learned are discussed. A Perceptual Paradox? Researchers who investigated how spatial memories are organised often suggest the existence of a two-system model: an egocentric model and an allocentric model (Mou, McNamara, Valiquette & Rump, 2004; Burgess, 2006; Rump & McNamara, 2007). These two models are very different implementations of the same basic mathematical model described above and therefore have different costs associated with their use. In particular, the former keeps track of the relationship between the self and all objects perceived. As one moves, one needs to constantly update all objects position in memory with respect to the viewer’s new position. The latter creates a global map of all objects perceived using a frame of reference independent of the viewer’s position. These researchers claimed that the former is best suited for organising information in a perceptual map while the latter is best for a cognitive map. However, little is said about how information encoded in an egocentric perceptual map is transferred into an allocentric cognitive map. If this is achieved via switching frame of reference, then the process is straightforward and from a mathematical standpoint, the two representations are considered equivalent. In this case, a perceptual map is a subset of a cognitive map and holds only the most recently perceived information. Researchers who investigated the nature of cognitive maps from studying resident’s memory of their environment (both adults and children) often emphasized that the map is fragmented, incomplete and imprecise (e.g. Lynch, 1960; Downs & Stea, 1973, Evans, 1980). This does not mean that the map is devoid of metric information but rather, one’s memory of such information is often found to be distorted systematically as a result of applying cognitive organizing principles (Tversky, 1992). Some well-known examples of these distortions include the regularization of turns and angles (Byrne, 1979), and over- and under- estimation of distances due to factors such as direction of travel (Lee, 1970), presence of barriers (Cohen & Weatherford, 1981), and others. More recent studies have also shown that metric Keywords: perceptual map; cognitive map; spatial layout; spatial cognition. Introduction How do humans integrate successive views to form a perceptual map? The latter is a representation of the spatial layout of surfaces/objects perceived in one’s immediate surroundings. That we have such a map is evident in that we do not immediately forget what is out of sight when we turn or move forward (see Glennerster, Hansard & Fitzgibbon (p.205, 2009) for a similar argument). However, researchers studying this problem from four different perspectives, namely how we represent our environmental knowledge (i.e. a cognitive map (Tolman, 1948; O’Keefe & Nadel, 1978)), what frame of references we use, how we see our world, and how robots create a map of their own world, have offered solutions which when taken together create a paradoxical situation. It is noted that because the problem lends itself to a straightforward mathematical solution whereby information in one view is transformed to their respective positions in the next view, much of the current studies implicitly or explicitly assume that a solution to this problem would involve such a step. This step is problematic when used to explain how humans integrate their views and the lack of an alternative method has hampered progress. In this paper, a new computational theory of human perceptual mapping is presented. It abandons the idea of integrating successive views to form a perceptual map. Instead, it argues that what is afforded in a view is an adequate description of the current spatial local environment and hence it does not need to be updated until one moves out of it. Only then, another view is added to the map. As a result, the map is composed of views selected at different times during one’s exploration of the environment.

[1]  D. E. Irwin,et al.  Eye movements and scene perception: Memory for things observed , 2002, Perception & psychophysics.

[2]  Ronald A. Rensink,et al.  Change blindness: past, present, and future , 2005, Trends in Cognitive Sciences.

[3]  D. Whitteridge Movements of the eyes R. H. S. Carpenter, Pion Ltd, London (1977), 420 pp., $27.00 , 1979, Neuroscience.

[4]  Wai-Kiang Yeap,et al.  Robotics and Cognitive Approaches to Spatial Mapping , 2010, Springer Tracts in Advanced Robotics.

[5]  M. Land Motion and vision: why animals move their eyes , 1999, Journal of Comparative Physiology A.

[6]  E. Tolman Cognitive maps in rats and men. , 1948, Psychological review.

[7]  Barbara Tversky,et al.  Distortions in cognitive maps , 1992 .

[8]  Wai-Kiang Yeap,et al.  Computing a Representation of the Local Environment , 1999, Artif. Intell..

[9]  R. Byrne,et al.  Memory for Urban Geography , 1979 .

[10]  Timothy P. McNamara,et al.  Updating in Models of Spatial Memory , 2006, Spatial Cognition.

[11]  R. Downs,et al.  Image and Environment: Cognitive Mapping and Spatial Behavior , 2017 .

[12]  R. Cohen,et al.  The effect of barriers on spatial representations. , 1981, Child development.

[13]  N. Burgess,et al.  Spatial memory: how egocentric and allocentric combine , 2006, Trends in Cognitive Sciences.

[14]  Wai-Kiang Yeap Towards a Computational Theory of Cognitive Maps , 1988, Artif. Intell..

[15]  Loong Fah Cheong,et al.  Visual space distortion , 1997, Biological Cybernetics.

[16]  Terence Lee Perceived Distance as a Function of Direction in the City , 1970 .

[17]  Bernhard Nebel,et al.  Spatial Cognition: Reasoning, Action, Interaction , 2010, KI - Künstliche Intelligenz.

[18]  Libor Preucil,et al.  European Robotics Symposium 2008 , 2008 .

[19]  R J Snowden Visual perception: Here's mud in your mind's eye , 1999, Current Biology.

[20]  Bodo Rosenhahn,et al.  Statistical and Geometrical Approaches to Visual Motion Analysis, International Dagstuhl Seminar, Dagstuhl Castle, Germany, July 13-18, 2008. Revised Papers , 2009, Lecture Notes in Computer Science.

[21]  E. Woolmington Image and Environment: Cognitive Mapping and Spatial Behaviour, R.M. Downs and D. Stea (eds), Aldine Publishing Company, Chicago, 1973. 160 x 240 mm., xxiv and 439 pages (with index), 36 tables, 117 figures. $15.00. , 1976 .

[22]  D. R. Montello,et al.  Spatial knowledge acquisition from direct experience in the environment: Individual differences in the development of metric knowledge and the integration of separately learned places , 2006, Cognitive Psychology.

[23]  R. Passingham The hippocampus as a cognitive map J. O'Keefe & L. Nadel, Oxford University Press, Oxford (1978). 570 pp., £25.00 , 1979, Neuroscience.

[24]  D. E. Irwin Integrating Information Across Saccadic Eye Movements , 1996 .

[25]  Kevin Lynch,et al.  The Image of the City , 1960 .

[26]  Andrew W. Fitzgibbon,et al.  View-Based Approaches to Spatial Representation in Human Vision , 2009, Statistical and Geometrical Approaches to Visual Motion Analysis.

[27]  Refractor Vision , 2000, The Lancet.

[28]  Wai-Kiang Yeap,et al.  How Albot0 Finds Its Way Home: A Novel Approach to Cognitive Mapping Using Robots , 2011, Top. Cogn. Sci..

[29]  Weimin Mou,et al.  Allocentric and egocentric updating of spatial memories. , 2004, Journal of experimental psychology. Learning, memory, and cognition.

[30]  Bruce Bridgeman,et al.  Processing Spatial Layout by Perception and Sensorimotor Interaction , 2008, Quarterly journal of experimental psychology.

[31]  H. Intraub The representation of visual scenes , 1997, Trends in Cognitive Sciences.

[32]  Petra Jansen-Osmann,et al.  Is Route Learning More Than Serial Learning? , 2008, Spatial Cogn. Comput..

[33]  J. O'Regan,et al.  Solving the "real" mysteries of visual perception: the world as an outside memory. , 1992, Canadian journal of psychology.

[34]  Bernhard Schölkopf,et al.  View-Based Cognitive Mapping and Path Planning , 1995, Adapt. Behav..