Encodji: encoding gaze data into emoji space for an amusing scanpath classification approach ;)

To this day, a variety of information has been obtained from human eye movements, which holds an imense potential to understand and classify cognitive processes and states - e.g., through scanpath classification. In this work, we explore the task of scanpath classification through a combination of unsupervised feature learning and convolutional neural networks. As an amusement factor, we use an Emoji space representation as feature space. This representation is achieved by training generative adversarial networks (GANs) for unpaired scanpath-to-Emoji translation with a cyclic loss. The resulting Emojis are then used to train a convolutional neural network for stimulus prediciton, showing an accuracy improvement of more than five percentual points compared to the same network trained using solely the scanpath data. As a side effect, we also obtain novel unique Emojis representing each unique scanpath. Our goal is to demonstrate the applicability and potential of unsupervised feature learning to scanpath classification in a humorous and entertaining way.

[1]  J. Piven,et al.  Visual Scanning of Faces in Autism , 2002, Journal of autism and developmental disorders.

[2]  Jean-Pierre Thibaut,et al.  An evaluation of scanpath-comparison and machine-learning classification algorithms used to study the dynamics of analogy making , 2016, Behavior Research Methods.

[3]  Trevor Darrell,et al.  Fully Convolutional Networks for Semantic Segmentation , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Stella X. Yu,et al.  Unsupervised Feature Learning via Non-parametric Instance Discrimination , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[5]  E. Gordon,et al.  Face to face: visual scanpath evidence for abnormal processing of facial expressions in social phobia , 2004, Psychiatry Research.

[6]  Alexei A. Efros,et al.  Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[7]  Tomas Pfister,et al.  Learning from Simulated and Unsupervised Images through Adversarial Training , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Antoine Coutrot,et al.  Scanpath modeling and classification with hidden Markov models , 2017, Behavior Research Methods.

[9]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  T. Crawford,et al.  How do radiologists do it? The influence of experience and training on searching for chest nodules. , 2006 .

[11]  Wolfgang Rosenstiel,et al.  SubsMatch 2.0: Scanpath comparison and classification based on subsequence frequencies , 2016, Behavior Research Methods.

[12]  Shiguang Shan,et al.  Duplex Generative Adversarial Network for Unsupervised Domain Adaptation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[13]  Leanne M Williams,et al.  Visual scanpaths to positive and negative facial emotions in an outpatient schizophrenia sample , 2002, Schizophrenia Research.

[14]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[15]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[16]  Enkelejda Kasneci,et al.  Automated Visual Scanpath Analysis Reveals the Expertise Level of Micro-neurosurgeons , 2015 .

[17]  Thomas C. Kübler,et al.  Driving with Glaucoma: Task Performance and Gaze Movements , 2015, Optometry and vision science : official publication of the American Academy of Optometry.

[18]  T. Loetscher,et al.  Eye Movements During Everyday Behavior Predict Personality Traits , 2018, Front. Hum. Neurosci..

[19]  Patrice Y. Simard,et al.  Best practices for convolutional neural networks applied to visual document analysis , 2003, Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings..

[20]  L. Stark,et al.  Evidence for a global scanpath strategy in viewing abstract compared with realistic images , 1995, Neuropsychologia.

[21]  Jan Theeuwes,et al.  ScanMatch: A novel method for comparing fixation sequences , 2010, Behavior research methods.

[22]  Roger Johansson,et al.  How task demands influence scanpath similarity in a sequential number-search task , 2018, Vision Research.

[23]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Trevor Darrell,et al.  Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.

[25]  Georg Gartner,et al.  Inferring user tasks in pedestrian navigation from eye movement data in real-world environments , 2018, Int. J. Geogr. Inf. Sci..

[26]  Javier Nogueras-Iso,et al.  A method for checking the quality of geographic metadata based on ISO 19157 , 2018, Int. J. Geogr. Inf. Sci..

[27]  Michael Burch,et al.  EyeMSA: exploring eye movement data with pairwise and multiple sequence alignment , 2018, ETRA.

[28]  Zhiwei Zhu,et al.  Real-time nonintrusive monitoring and prediction of driver fatigue , 2004, IEEE Transactions on Vehicular Technology.

[29]  Alan Kingstone,et al.  A comparison of scanpath comparison methods , 2014, Behavior Research Methods.

[30]  C. J. Ravesloot,et al.  How visual search relates to visual diagnostic performance: a narrative systematic review of eye-tracking research in radiology , 2017, Advances in health sciences education : theory and practice.

[31]  Stephen L Macknik,et al.  Highly Informative Natural Scene Regions Increase Microsaccade Production during Visual Scanning , 2014, The Journal of Neuroscience.

[32]  Haogang Zhu,et al.  What's on TV? Detecting age-related neurodegenerative eye disease using eye movement scanpaths , 2014, Front. Aging Neurosci..

[33]  Andrew L. Kun,et al.  Estimating cognitive load using remote eye tracking in a driving simulator , 2010, ETRA.

[34]  Tianyi Zhang,et al.  How Old Do You Look? Inferring Your Age from Your Gaze , 2018, 2018 25th IEEE International Conference on Image Processing (ICIP).

[35]  S. Srihari Mixture Density Networks , 1994 .

[36]  Elizabeth A. Krupinski,et al.  Art and authenticity: Behavioral and eye-movement analyses. , 2015 .

[37]  Markku Tukiainen,et al.  Gaze behaviour of expert and novice microneurosurgeons differs during observations of tumor removal recordings , 2012, ETRA '12.

[38]  L. Stark,et al.  Spontaneous Eye Movements During Visual Imagery Reflect the Content of the Visual Scene , 1997, Journal of Cognitive Neuroscience.

[39]  W. Rosenstiel,et al.  Driving with Binocular Visual Field Loss? A Study on a Supervised On-Road Parcours with Simultaneous Eye and Head Tracking , 2014, PloS one.

[40]  Katharina Scheiter,et al.  Scanpath comparison in medical image reading skills of dental students: distinguishing stages of expertise development , 2018, ETRA.

[41]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[42]  Xoana G. Troncoso,et al.  Saccades and microsaccades during visual fixation, exploration, and search: foundations for a common saccadic generator. , 2008, Journal of vision.

[43]  V. Gallese,et al.  When Art Moves the Eyes: A Behavioral and Eye-Tracking Study , 2012, PloS one.

[44]  拓海 杉山,et al.  “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”の学習報告 , 2017 .

[45]  Wolfgang Rosenstiel,et al.  Online Recognition of Driver-Activity Based on Visual Scanpath Classification , 2017, IEEE Intelligent Transportation Systems Magazine.

[46]  Neil D. B. Bruce,et al.  Predicting task from eye movements: On the importance of spatial distribution, dynamics, and image features , 2016, Neurocomputing.