Video-based eye trackers produce an output video showing where a subject is looking, the subject's point-of-regard (POR), for each frame of a video of the scene. Fixation-identification algorithms simplify the long list of POR data into a more manageable set of data, especially for further analysis, by grouping PORs into fixations. Most current fixation-identification algorithms assume that the POR data are defined in static two-dimensional scene images and only use these raw POR data to identify fixations. The applicability of these algorithms to gaze data in dynamic scene videos is largely unexplored. We implemented a simple velocity-based, duration-sensitive fixation-identification algorithm and compared its performance to results obtained by three experienced users manually coding the eye tracking data displayed within the scene video such that these manual coders had knowledge of the scene motion. We performed this comparison for eye tracking data collected during two different tasks involving different types of scene motion. These two tasks included a subject walking around a building for about 100 seconds (Task 1) and a seated subject viewing a computer animation (approximately 90 seconds long, Task 2). It took our manual coders on average 75 minutes (stdev = 28) and 80 minutes (17) to code results from the first and second tasks, respectively. The automatic fixation-identification algorithm, implemented in MATLAB and run on an Apple 2.16 GHz MacBook, produced results in 0.26 seconds for Task 1 and 0.21 seconds for Task 2. For the first task (walking), the average percent difference among the three human manual coders was 9% (3.5) and the average percent difference between the automatically generated results and the three coders was 11% (2.0). For the second task (animation), the average percent difference among the three human coders was 4% (0.75) and the average percent difference between the automatically generated results and the three coders was 5% (0.9).
[1]
K. Rayner.
Eye movements and visual cognition : scene perception and reading
,
1992
.
[2]
Laurent Itti,et al.
Computational mechanisms for gaze direction in interactive visual environments
,
2006,
ETRA.
[3]
Jeff B. Pelz,et al.
Building a lightweight eyetracking headgear
,
2004,
ETRA.
[4]
Eli Peli,et al.
Where people look when watching movies: Do all viewers look at the same place?
,
2007,
Comput. Biol. Medicine.
[5]
Laurent Itti,et al.
Causal saliency effects during natural vision
,
2006,
ETRA.
[6]
Joseph H. Goldberg,et al.
Identifying fixations and saccades in eye-tracking protocols
,
2000,
ETRA.
[7]
M. Just,et al.
Eye fixations and cognitive processes
,
1976,
Cognitive Psychology.