Power-efficient and shift-robust eye-tracking sensor for portable VR headsets

Photosensor oculography (PSOG) is a promising solution for reducing the computational requirements of eye tracking sensors in wireless virtual and augmented reality platforms. This paper proposes a novel machine learning-based solution for addressing the known performance degradation of PSOG devices in the presence of sensor shifts. Namely, we introduce a convolutional neural network model capable of providing shift-robust end-to-end gaze estimates from the PSOG array output. Moreover, we propose a transfer-learning strategy for reducing model training time. Using a simulated workflow with improved realism, we show that the proposed convolutional model offers improved accuracy over a previously considered multilayer perceptron approach. In addition, we demonstrate that the transfer of initialization weights from pre-trained models can substantially reduce training time for new users. In the end, we provide the discussion regarding the design trade-offs between accuracy, training time, and power consumption among the considered models.

[1]  Lee Friedman,et al.  Method to assess the temporal persistence of potential biometric features: Application to oculomotor, gait, face and brain structure databases , 2016, PloS one.

[2]  Qiang Liu,et al.  Ultra-Low Power Gaze Tracking for Virtual Reality , 2017, SenSys.

[3]  Timothy Dozat,et al.  Incorporating Nesterov Momentum into Adam , 2016 .

[4]  Chris Harrison,et al.  EyeSpyVR: Interactive Eye Sensing Using Off-the-Shelf, Smartphone-Based VR Headsets , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[5]  Ioannis Rigas,et al.  Hybrid PS-V Technique: A Novel Sensor Fusion Approach for Fast Mobile Eye-Tracking With Sensor-Shift Aware Correction , 2017, IEEE Sensors Journal.

[6]  Ioannis Rigas,et al.  Photosensor Oculography: Survey and Parametric Analysis of Designs Using Model-Based Simulation , 2017, IEEE Transactions on Human-Machine Systems.

[7]  Rich Caruana,et al.  Overfitting in Neural Nets: Backpropagation, Conjugate Gradient, and Early Stopping , 2000, NIPS.

[8]  Desney S. Tan,et al.  Foveated 3D graphics , 2012, ACM Trans. Graph..

[9]  Alexander LeNail,et al.  NN-SVG: Publication-Ready Neural Network Architecture Schematics , 2019, J. Open Source Softw..

[10]  Yoshua Bengio,et al.  Practical Recommendations for Gradient-Based Training of Deep Architectures , 2012, Neural Networks: Tricks of the Trade.

[11]  J. Hess,et al.  Analysis of variance , 2018, Transfusion.

[12]  V. Guillemin,et al.  LXXX Photoelectric Nystagmography , 1951 .

[13]  Oleg V. Komogortsev,et al.  Study of Additional Eye-Related Features for Future Eye-Tracking Techniques , 2017, CHI Extended Abstracts.

[14]  Oleg V. Komogortsev,et al.  Making stand-alone PS-OG technology tolerant to the equipment shifts , 2018, PETMEI@ETRA.