Machine learning-based classification of viewing behavior using a wide range of statistical oculomotor features

Since the seminal work of Yarbus, multiple studies have demonstrated the influence of task-set on oculomotor behavior and the current cognitive state. In more recent years, this field of research has expanded by evaluating the costs of abruptly switching between such different tasks. At the same time, the field of classifying oculomotor behavior has been moving toward more advanced, data-driven methods of decoding data. For the current study, we used a large dataset compiled over multiple experiments and implemented separate state-of-the-art machine learning methods for decoding both cognitive state and task-switching. We found that, by extracting a wide range of oculomotor features, we were able to implement robust classifier models for decoding both cognitive state and task-switching. Our decoding performance highlights the feasibility of this approach, even invariant of image statistics. Additionally, we present a feature ranking for both models, indicating the relative magnitude of different oculomotor features for both classifiers. These rankings indicate a separate set of important predictors for decoding each task, respectively. Finally, we discuss the implications of the current approach related to interpreting the decoding results.

[1]  T. Carlson,et al.  High temporal resolution decoding of object position and category. , 2011, Journal of vision.

[2]  James J. Clark,et al.  An inverse Yarbus process: Predicting observers’ task from eye movement patterns , 2014, Vision Research.

[3]  Ron Kohavi,et al.  A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection , 1995, IJCAI.

[4]  Tanya Wen,et al.  Using functional magnetic resonance imaging to explore the flashed face distortion effect. , 2014, Journal of vision.

[5]  Casimir A. Kulikowski,et al.  Computer Systems That Learn: Classification and Prediction Methods from Statistics, Neural Nets, Machine Learning and Expert Systems , 1990 .

[6]  Michelle R. Greene,et al.  Reconsidering Yarbus: A failure to predict observers’ task from eye movement patterns , 2012, Vision Research.

[7]  L. Itti,et al.  High-throughput classification of clinical populations from natural viewing eye movements , 2012, Journal of Neurology.

[8]  A. L. Yarbus Eye Movements During Perception of Complex Objects , 1967 .

[9]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[10]  J. Wolfe,et al.  The role of memory for visual search in scenes , 2015, Annals of the New York Academy of Sciences.

[11]  Andrew Hollingworth,et al.  Eye Movements During Scene Viewing: An Overview , 1998 .

[12]  Pat Langley,et al.  Selection of Relevant Features and Examples in Machine Learning , 1997, Artif. Intell..

[13]  Wes McKinney,et al.  Data Structures for Statistical Computing in Python , 2010, SciPy.

[14]  Alexander C. Schütz,et al.  Eye movements and perception: a selective review. , 2011, Journal of vision.

[15]  Joy J. Geng,et al.  Pupil diameter reflects uncertainty in attentional selection during visual search , 2015, Front. Hum. Neurosci..

[16]  Travis E. Oliphant,et al.  Python for Scientific Computing , 2007, Computing in Science & Engineering.

[17]  John M. Henderson,et al.  Predicting Cognitive State from Eye Movements , 2013, PloS one.

[18]  Michael D. Dodd,et al.  Effects of task and task-switching on temporal inhibition of return, facilitation of return, and saccadic momentum during scene viewing. , 2015, Journal of experimental psychology. Human perception and performance.

[19]  L. Itti,et al.  Defending Yarbus: eye movements reveal observers' task. , 2014, Journal of vision.

[20]  Taylor R. Hayes,et al.  Mapping and correcting the influence of gaze position on pupil size measurements , 2015, Behavior Research Methods.

[21]  Eric Jones,et al.  SciPy: Open Source Scientific Tools for Python , 2001 .

[22]  Michael D. Dodd,et al.  A Generative Model of Cognitive State from Task and Eye Movements , 2018, Cognitive Computation.

[23]  M. Bradley,et al.  The pupil as a measure of emotional arousal and autonomic activation. , 2008, Psychophysiology.

[24]  Sotiris B. Kotsiantis,et al.  Supervised Machine Learning: A Review of Classification Techniques , 2007, Informatica.

[25]  Nando de Freitas,et al.  Taking the Human Out of the Loop: A Review of Bayesian Optimization , 2016, Proceedings of the IEEE.

[26]  Mieke Donk,et al.  Conditional control in visual selection , 2017, Attention, perception & psychophysics.

[27]  Michael D. Dodd,et al.  Examining the influence of task set on eye movements and fixations. , 2011, Journal of vision.

[28]  Oleg V. Komogortsev,et al.  Using machine learning to detect events in eye-tracking data , 2018, Behavior research methods.

[29]  G. D. Logan Task Switching , 2022 .

[30]  Gaël Varoquaux,et al.  The NumPy Array: A Structure for Efficient Numerical Computation , 2011, Computing in Science & Engineering.

[31]  C. Koch,et al.  Computational modelling of visual attention , 2001, Nature Reviews Neuroscience.

[32]  Michael D. Dodd,et al.  Human classifier: Observers can deduce task solely from eye movements , 2017, Attention, Perception, & Psychophysics.