Gaze-Assisted User Intention Prediction for Initial Delay Reduction in Web Video Access

Despite the remarkable improvement of hardware and network technology, the inevitable delay from a user's command action to a system response is still one of the most crucial influence factors in user experiences (UXs). Especially for a web video service, an initial delay from click action to video start has significant influences on the quality of experience (QoE). The initial delay of a system can be minimized by preparing execution based on predicted user's intention prior to actual command action. The introduction of the sequential and concurrent flow of resources in human cognition and behavior can significantly improve the accuracy and preparation time for intention prediction. This paper introduces a threaded interaction model and applies it to user intention prediction for initial delay reduction in web video access. The proposed technique consists of a candidate selection module, a decision module and a preparation module that prefetches and preloads the web video data before a user's click action. The candidate selection module selects candidates in the web page using proximity calculation around a cursor. Meanwhile, the decision module computes the possibility of actual click action based on the cursor-gaze relationship. The preparation activates the prefetching for the selected candidates when the click possibility exceeds a certain limit in the decision module. Experimental results show a 92% hit-ratio, 0.5-s initial delay on average and 1.5-s worst initial delay, which is much less than a user's tolerable limit in web video access, demonstrating significant improvement of accuracy and advance time in intention prediction by introducing the proposed threaded interaction model.

[1]  Simon J. Godsill,et al.  User Target Intention Recognition from Cursor Position Using Kalman Filter , 2013, HCI.

[2]  Da Young Ju,et al.  Data Preloading Technique using Intention Prediction , 2014, HCI.

[3]  Hiroshi Sato,et al.  MobiGaze: development of a gaze interface for handheld mobile devices , 2010, CHI EA '10.

[4]  Dominik Schmidt,et al.  Eye Pull, Eye Push: Moving Objects between Large Screens and Personal Devices with Gaze and Touch , 2013, INTERACT.

[5]  Peter Corcoran To Gaze with Undimmed Eyes on All Darkness [IP Corner] , 2015, IEEE Consumer Electronics Magazine.

[6]  Robert J. K. Jacob,et al.  What you look at is what you get: eye movement-based interaction techniques , 1990, CHI '90.

[7]  Leonard Kleinrock,et al.  Web prefetching in a mobile environment , 1998, IEEE Wirel. Commun..

[8]  Päivi Majaranta,et al.  Twenty years of eye typing: systems and design issues , 2002, ETRA.

[9]  Patrick Langdon,et al.  Multimodal Intelligent Eye-Gaze Tracking System , 2015, Int. J. Hum. Comput. Interact..

[10]  Jacob O. Wobbrock,et al.  Mouse pointing endpoint prediction using kinematic template matching , 2014, CHI.

[11]  Simon J. Godsill,et al.  Intent Recognition Using Neural Networks and Kalman Filters , 2013, CHI-KDD.

[12]  Gabriella Pasi,et al.  Human-Computer Interaction and Knowledge Discovery in Complex, Unstructured, Big Data , 2013, Lecture Notes in Computer Science.

[13]  Andreas Bulling,et al.  Combining gaze with manual interaction to extend physical reach , 2011, PETMEI '11.

[14]  Anirban Kundu,et al.  Dynamic Web Prediction Using Asynchronous Mouse Activity , 2012 .

[15]  Pradipta Biswas,et al.  Effect of Road Conditions on Gaze-Control Interface in an Automotive Environment , 2015, HCI.

[16]  Jiahui Liu,et al.  Personalized news recommendation based on click behavior , 2010, IUI '10.

[17]  Matthias Vogelgesang,et al.  Multimodal integration of natural gaze behavior for intention recognition during object manipulation , 2009, ICMI-MLMI '09.

[18]  Margherita Antona,et al.  Universal Access in Human-Computer Interaction. Design Methods, Tools, and Interaction Techniques for eInclusion , 2013, Lecture Notes in Computer Science.

[19]  Leonard Kleinrock,et al.  An adaptive network prefetch scheme , 1998, IEEE J. Sel. Areas Commun..

[20]  Fiona Fui-Hoon Nah,et al.  A study on tolerable waiting time: how long are Web users willing to wait? , 2004, AMCIS.

[21]  Hans-Werner Gellersen,et al.  Multimodal recognition of reading activity in transit using body-worn sensors , 2012, TAP.

[22]  Sebastian Maier,et al.  Eye gaze assisted human-computer interaction in a hand gesture controlled multi-display environment , 2012, Gaze-In '12.

[23]  Roman Bednarik,et al.  What do you want to do next: a novel approach for intent prediction in gaze-based interaction , 2012, ETRA.

[24]  Michael S. Borella,et al.  The effect of network delay and media on user perceptions of web resources , 2000, Behav. Inf. Technol..

[25]  John R. Anderson,et al.  Intelligent gaze-added interfaces , 2000, CHI.

[26]  Hyokyung Bahn,et al.  Channel reordering and prefetching schemes for efficient IPTV channel navigation , 2010, IEEE Transactions on Consumer Electronics.

[27]  John Paulin Hansen,et al.  Single stroke gaze gestures , 2009, CHI Extended Abstracts.

[28]  Mario Fritz,et al.  Prediction of search targets from fixations in open-world settings , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  Raimund Dachselt,et al.  Look & touch: gaze-supported target acquisition , 2012, CHI.

[30]  Hans-Werner Gellersen,et al.  EyeContext: recognition of high-level contextual cues from human visual behaviour , 2013, CHI.

[31]  Gerhard Tröster,et al.  Eye Movement Analysis for Activity Recognition Using Electrooculography , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  Niels Taatgen,et al.  The Multitasking Mind , 2010, Oxford series on cognitive models and architectures.

[33]  Krishna Bharat,et al.  Making computers easier for older adults to use: area cursors and sticky icons , 1997, CHI.

[34]  Atsuo Murata,et al.  Improvement of Pointing Time by Predicting Targets in Pointing With a PC Mouse , 1998, Int. J. Hum. Comput. Interact..

[35]  Dario D. Salvucci,et al.  Threaded cognition: an integrated theory of concurrent multitasking. , 2008, Psychological review.

[36]  Olivier Chapuis,et al.  DynaSpot: speed-dependent area cursor , 2009, CHI.

[37]  Andreas Bulling,et al.  Cognition-Aware Computing , 2014, IEEE Pervasive Computing.

[38]  Ramesh K. Sitaraman,et al.  Video Stream Quality Impacts Viewer Behavior: Inferring Causality Using Quasi-Experimental Designs , 2012, IEEE/ACM Transactions on Networking.

[39]  Rynson W. H. Lau,et al.  The implicit fan cursor: a velocity dependent area cursor , 2014, CHI.

[40]  Anirban Kundu,et al.  A new approach in dynamic prediction for user based web page crawling , 2010, MEDES.

[41]  Tovi Grossman,et al.  The bubble cursor: enhancing target acquisition by dynamic resizing of the cursor's activation area , 2005, CHI.

[42]  Joe Marini,et al.  Document Object Model , 2002, Encyclopedia of GIS.

[43]  Janna Protzak,et al.  A Passive Brain-Computer Interface for Supporting Gaze-Based Human-Machine Interaction , 2013, HCI.