Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities

The study of gaze behavior has primarily been constrained to controlled environments in which the head is fixed. Consequently, little effort has been invested in the development of algorithms for the categorization of gaze events (e.g. fixations, pursuits, saccade, gaze shifts) while the head is free, and thus contributes to the velocity signals upon which classification algorithms typically operate. Our approach was to collect a novel, naturalistic, and multimodal dataset of eye + head movements when subjects performed everyday tasks while wearing a mobile eye tracker equipped with an inertial measurement unit and a 3D stereo camera. This Gaze-in-the-Wild dataset (GW) includes eye + head rotational velocities (deg/s), infrared eye images and scene imagery (RGB + D). A portion was labelled by coders into gaze motion events with a mutual agreement of 0.74 sample based Cohen’s κ. This labelled data was used to train and evaluate two machine learning algorithms, Random Forest and a Recurrent Neural Network model, for gaze event classification. Assessment involved the application of established and novel event based performance metrics. Classifiers achieve ~87% human performance in detecting fixations and saccades but fall short (50%) on detecting pursuit movements. Moreover, pursuit classification is far worse in the absence of head movement information. A subsequent analysis of feature significance in our best performing model revealed that classification can be done using only the magnitudes of eye and head movements, potentially removing the need for calibration between the head and eye tracking systems. The GW dataset, trained classifiers and evaluation metrics will be made publicly available with the intention of facilitating growth in the emerging area of head-free gaze event classification.

[1]  H. Collewijn,et al.  The function of visual search and memory in sequential looking tasks , 1995, Vision Research.

[2]  Marcus Nyström,et al.  Is human classification by experienced untrained observers a gold standard in fixation detection? , 2017, Behavior research methods.

[3]  R.S. Allison,et al.  Combined head and eye tracking system for dynamic testing of the vestibular system , 1996, IEEE Transactions on Biomedical Engineering.

[4]  D. Ballard,et al.  Eye movements in natural behavior , 2005, Trends in Cognitive Sciences.

[5]  Oleg V Komogortsev,et al.  Automated classification and scoring of smooth pursuit eye movements in the presence of fixations and saccades , 2013, Behavior research methods.

[6]  Andreas Bulling,et al.  Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction , 2014, UbiComp Adjunct.

[7]  Sébastien Ourselin,et al.  Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations , 2017, DLMIA/ML-CDS@MICCAI.

[8]  Kenneth Holmqvist,et al.  End-to-end eye-movement event detection using deep neural networks , 2017 .

[9]  Richard Kempter,et al.  State-dependencies of learning across brain scales , 2015, Front. Comput. Neurosci..

[10]  Rochelle Ackerley,et al.  The interaction of visual, vestibular and extra-retinal mechanisms in the control of head and gaze during head-free pursuit , 2011, The Journal of physiology.

[11]  Gang Luo,et al.  Mobile gaze tracking system for outdoor walking behavioral studies , 2016, Journal of vision.

[12]  Luis F. Chaparro Chapter 2 – Continuous-time Systems , 2015 .

[13]  Yoshua Bengio,et al.  Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.

[14]  A. A. Skavenski,et al.  Quality of retinal image stabilization during small natural and artificial body rotations in man , 1979, Vision Research.

[15]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[16]  Alan Kennedy,et al.  Book Review: Eye Tracking: A Comprehensive Guide to Methods and Measures , 2016, Quarterly journal of experimental psychology.

[17]  Otto Lappi,et al.  A new and general approach to signal denoising and eye movement classification based on segmented linear regression , 2017, Scientific Reports.

[18]  M. Hayhoe,et al.  In what ways do eye movements contribute to everyday activities? , 2001, Vision Research.

[19]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .

[20]  Thiago Santini,et al.  Bayesian identification of fixations, saccades, and smooth pursuits , 2015, ETRA.

[21]  Otto Lappi,et al.  Eye movements in the wild: Oculomotor control, gaze behavior & frames of reference , 2016, Neuroscience & Biobehavioral Reviews.

[22]  S. Moss One algorithm to rule them all , 2017 .

[23]  Kenneth Holmqvist,et al.  gazeNet: End-to-end eye-movement event detection with deep neural networks , 2018, Behavior Research Methods.

[24]  Gunnar Blohm,et al.  Catch-up saccades in head-unrestrained conditions reveal that saccade amplitude is corrected using an internal model of target movement. , 2014, Journal of vision.

[25]  Uwe D. Reichel,et al.  A dialect distance metric based on string and temporal alignment , 2013 .

[26]  Jeff B. Pelz,et al.  Predictive eye movements in natural vision , 2011, Experimental Brain Research.

[27]  J. Douglas Crawford,et al.  A kinematic model for 3-D head-free gaze-shifts , 2015, Front. Comput. Neurosci..

[28]  Peter König,et al.  Human eye-head co-ordination in natural exploration , 2007, Network.

[29]  Marcus Nyström,et al.  Head movement compensation and multi-modal event detection in eye-tracking data for unconstrained head movements , 2016, Journal of Neuroscience Methods.

[30]  Ignace T. C. Hooge,et al.  Is the eye-movement field confused about fixations and saccades? A survey among 124 researchers , 2018, Royal Society Open Science.

[31]  G. Barnes,et al.  Visual-vestibular interaction in the control of head and eye movement: The role of visual feedback and predictive mechanisms , 1993, Progress in neurobiology.

[32]  Ioannis Agtzidis,et al.  In the pursuit of (ground) truth: a hand-labelling tool for eye movements recorded during dynamic scene viewing , 2016, 2016 IEEE Second Workshop on Eye Tracking and Visualization (ETVIS).

[33]  Michael F. Land,et al.  The human eye movement repertoire , 2009 .

[34]  Mary M. Hayhoe,et al.  Gaze and the Control of Foot Placement When Walking in Natural Terrain , 2018, Current Biology.

[35]  Luis F. Chaparro Continuous-Time Systems , 2011 .

[36]  Emily A. Cooper,et al.  Stereopsis is adaptive for the natural environment , 2015, Science Advances.

[37]  Karen M. Evans,et al.  Ego-motion compensation improves fixation detection in wearable eye tracking , 2012, ETRA.

[38]  Dora E Angelaki,et al.  Three-dimensional ocular kinematics during eccentric rotations: evidence for functional rather than mechanical constraints. , 2003, Journal of neurophysiology.

[39]  Dora E Angelaki,et al.  Eyes on target: what neurons must do for the vestibuloocular reflex during linear motion. , 2004, Journal of neurophysiology.

[40]  E. Freedman Coordination of the eyes and head during visual orienting , 2008, Experimental Brain Research.

[41]  B. Abernethy,et al.  The Head Tracks and Gaze Predicts: How the World’s Best Batters Hit a Ball , 2013, PloS one.

[42]  Ahmet Alkan,et al.  Automatic seizure detection in EEG using logistic regression and artificial neural network , 2005, Journal of Neuroscience Methods.

[43]  Luis E. Ortiz,et al.  Depth Data Error Modeling of the ZED 3D Vision Sensor from Stereolabs , 2018, ELCVIA Electronic Letters on Computer Vision and Image Analysis.

[44]  J. E. Lieberman,et al.  Frequency limitations of the two-point central difference differentiation algorithm , 1982, Biological Cybernetics.

[45]  G R Barnes,et al.  Head‐free pursuit in the human of a visual target moving in a pseudo‐random manner. , 1989, The Journal of physiology.

[46]  G. Barnes Vestibulo‐ocular function during co‐ordinated head and eye movements to acquire visual targets. , 1979, The Journal of physiology.

[47]  M. Mustari,et al.  Optokinetic Eye Movements , 2009 .

[48]  Brett R Fajen,et al.  Visual control of foot placement when walking over complex terrain , 2014, Journal of experimental psychology. Human perception and performance.

[49]  B. Tatler,et al.  Yarbus, eye movements, and vision , 2010, i-Perception.

[50]  Heikki Huttunen,et al.  Polyphonic sound event detection using multi label deep neural networks , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[51]  Oleg V. Komogortsev,et al.  Using machine learning to detect events in eye-tracking data , 2018, Behavior research methods.

[52]  Reynold J. Bailey,et al.  Novel apparatus for investigation of eye movements when walking in the presence of 3D projected obstacles , 2016, ETRA.

[53]  Andreas Bulling,et al.  End-to-End Eye Movement Detection Using Convolutional Neural Networks , 2016, ArXiv.

[54]  B. L. Zuber,et al.  Frequency characteristics of the saccadic eye movement. , 1968, Biophysical journal.

[55]  Andrew Zisserman,et al.  Multiple View Geometry in Computer Vision (2nd ed) , 2003 .

[56]  David M. W. Powers,et al.  The Problem with Kappa , 2012, EACL.

[57]  G M Gauthier,et al.  Eye-head movement coordination: vestibulo-ocular reflex suppression with head-fixed target fixation. , 1991, Journal of vestibular research : equilibrium & orientation.

[58]  D. Hubel,et al.  The role of fixational eye movements in visual perception , 2004, Nature Reviews Neuroscience.

[59]  Pierre Morel,et al.  Gramm: grammar of graphics plotting in Matlab , 2018, J. Open Source Softw..

[60]  Yu Fang,et al.  Eye-Head Coordination for Visual Cognitive Processing , 2015, PloS one.

[61]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[62]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.