RIT-Eyes: Rendering of near-eye images for eye-tracking applications

Deep neural networks for video-based eye tracking have demonstrated resilience to noisy environments, stray reflections, and low resolution. However, to train these networks, a large number of manually annotated images are required. To alleviate the cumbersome process of manual labeling, computer graphics rendering is employed to automatically generate a large corpus of annotated eye images under various conditions. In this work, we introduce a synthetic eye image generation platform that improves upon previous work by adding features such as an active deformable iris, an aspherical cornea, retinal retro-reflection, gaze-coordinated eye-lid deformations, and blinks. To demonstrate the utility of our platform, we render images reflecting the represented gaze distributions inherent in two publicly available datasets, NVGaze and OpenEDS. We also report on the performance of two semantic segmentation architectures (SegNet and RITnet) trained on rendered images and tested on the original datasets.

[1]  Reynold J. Bailey,et al.  RITnet: Real-time Semantic Segmentation of the Eye for Gaze Tracking , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[2]  Seyed-Ahmad Ahmadi,et al.  DeepVOG: Open-source pupil segmentation and gaze estimation in neuroscience using deep learning , 2019, Journal of Neuroscience Methods.

[3]  Thiago Santini,et al.  Improving real-time CNN-based pupil detection through domain-specific data augmentation , 2019, ETRA.

[4]  Christopher Kanan,et al.  Gaze-in-wild: A dataset for studying eye and head coordination in everyday activities , 2019, Scientific Reports.

[5]  Jan Kautz,et al.  Few-Shot Adaptive Gaze Estimation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[6]  Joohwan Kim,et al.  NVGaze: An Anatomically-Informed Dataset for Low-Latency, Near-Eye Gaze Estimation , 2019, CHI.

[7]  Derek Bradley,et al.  Practical Person‐Specific Eye Rigging , 2019, Comput. Graph. Forum.

[8]  Gregory Hughes,et al.  OpenEDS: Open Eye Dataset , 2019, ArXiv.

[9]  Ioannis Agtzidis,et al.  360-degree Video Gaze Behaviour: A Ground-Truth Data Set and a Classification Algorithm for Eye Movements , 2019, ACM Multimedia.

[10]  Jose Dolz,et al.  Boundary loss for highly unbalanced segmentation , 2018, MIDL.

[11]  Francisco Javier Vera-Olmos,et al.  DeepEye: Deep convolutional network for pupil detection in real environments , 2018, Integr. Comput. Aided Eng..

[12]  Andrew Zisserman,et al.  Turning a Blind Eye: Explicit Removal of Biases and Variation from Deep Neural Network Embeddings , 2018, ECCV Workshops.

[13]  Otmar Hilliges,et al.  Deep Pictorial Gaze Estimation , 2018, ECCV.

[14]  Erik Lindén,et al.  Learning to Personalize in Appearance-Based Gaze Tracking , 2018, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[15]  Andreas Bulling,et al.  A novel approach to single camera, glint-free 3D eye model fitting including corneal refraction , 2018, ETRA.

[16]  Otmar Hilliges,et al.  Learning to find eye region landmarks for remote gaze estimation in unconstrained settings , 2018, ETRA.

[17]  Wolfgang Rosenstiel,et al.  PupilNet v2.0: Convolutional Neural Networks for CPU based real time Robust Pupil Detection , 2017, ArXiv.

[18]  Sébastien Ourselin,et al.  Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations , 2017, DLMIA/ML-CDS@MICCAI.

[19]  Tomas Pfister,et al.  Learning from Simulated and Unsupervised Images through Adversarial Training , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Derek Bradley,et al.  Lightweight eye capture using a parametric model , 2016, ACM Trans. Graph..

[21]  Jia Deng,et al.  Stacked Hourglass Networks for Human Pose Estimation , 2016, ECCV.

[22]  Peter Robinson,et al.  Learning an appearance-based gaze estimator from one million synthesised images , 2016, ETRA.

[23]  Larry N Thibos,et al.  Optical models of the human eye , 2016, Clinical & experimental optometry.

[24]  Roberto Cipolla,et al.  SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  J. Meunier,et al.  Corneal Shape, Volume, and Interocular Symmetry: Parameters to Optimize the Design of Biosynthetic Corneal Substitutes. , 2015, Investigative ophthalmology & visual science.

[26]  Peter Robinson,et al.  Rendering of Eyes for Eye-Shape Registration and Gaze Estimation , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[27]  Mario Fritz,et al.  Appearance-based gaze estimation in the wild , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[29]  Derek Bradley,et al.  High-quality capture of eyes , 2014, ACM Trans. Graph..

[30]  Andreas Bulling,et al.  Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction , 2014, UbiComp Adjunct.

[31]  Neil A. Dodgson,et al.  Rendering synthetic ground truth images for eye tracker evaluation , 2014, ETRA.

[32]  Alan C. Bovik,et al.  The Essential Guide to Image Processing , 2009, J. Electronic Imaging.

[33]  Thomas Martinetz,et al.  A software framework for simulating eye trackers , 2008, ETRA.

[34]  Carlos Hitoshi Morimoto,et al.  Eye gaze tracking techniques for interactive applications , 2005, Comput. Vis. Image Underst..

[35]  Myron Flickner,et al.  Differences in the infrared bright pupil response of human eyes , 2002, ETRA.

[36]  Brian D. Davison,et al.  Learning to personalize , 2000, CACM.

[37]  V. Bapat,et al.  Turning a blind eye , 1997, The Lancet.

[38]  Neil Dodgson,et al.  A fully-automatic , temporal approach to single camera , glint-free 3 D eye model fitting , 2013 .

[39]  C. Peota Novel approach. , 2011, Minnesota medicine.

[40]  J. Daugman How iris recognition works , 2004, IEEE Transactions on Circuits and Systems for Video Technology.