Toward A Pervasive Gaze-Contingent Assistance System: Attention and Context-Awareness in Augmented Reality

Mobile devices with high-speed connectivity provide us with access to gigabytes of high resolution images, videos, and graphics. For instance, a head-worn display can be used to augment the real view with digitized visual information (Figure 1). Eye tracking helps us to understand how we process visual information and it allows us to develop gaze-enabled interactive systems. For instance, foveated gaze-contingent displays (GCDs) dynamically adjust the level of detail according to the user’s point-of-interest. We propose that GCDs should take users’ attention and cognitive load into account, augment their vision with contextual information and provide personalized assistance in solving visual tasks. Grounded on existing literature, we identified several research questions that need to be discussed before developing such displays.

[1]  Andreas Bulling,et al.  Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction , 2014, UbiComp Adjunct.

[2]  Ernst Niebur,et al.  Variable-Resolution Displays: A Theoretical, Practical, and Behavioral Evaluation , 2002, Hum. Factors.

[3]  Yixiang Lim,et al.  Cognitive Human-Machine Interfaces and Interactions for Unmanned Aircraft , 2018, J. Intell. Robotic Syst..

[4]  Martin Raubal,et al.  Measuring Cognitive Load for Map Tasks Through Pupil Diameter , 2016, GIScience.

[5]  Joohwan Kim,et al.  Foveated AR , 2019, ACM Trans. Graph..

[6]  Zhou Wang,et al.  Foveated wavelet image quality index , 2001, Optics + Photonics.

[7]  Kiyoshi Kiyokawa,et al.  Virtual and Augmented Reality on the 5G Highway , 2017, J. Inf. Process..

[8]  Holger Regenbrecht,et al.  Towards Pervasive Augmented Reality: Context-Awareness in Augmented Reality , 2017, IEEE Transactions on Visualization and Computer Graphics.

[9]  Arzu Çöltekin,et al.  Measured and perceived visual complexity: a comparative study among three online map providers , 2018 .

[10]  Desney S. Tan,et al.  Foveated 3D graphics , 2012, ACM Trans. Graph..

[11]  Andrew T. Duchowski,et al.  Hybrid image-/model-based gaze-contingent rendering , 2007, TAP.

[12]  Alan C. Bovik,et al.  Foveated Image and Video Processing and Search , 2014 .

[13]  Zhou Wang,et al.  Embedded foveation image coding , 2001, IEEE Trans. Image Process..

[14]  D. Pelli,et al.  Crowding is unlike ordinary masking: distinguishing feature integration from detection. , 2004, Journal of vision.

[15]  Alexander Toet,et al.  Gaze directed displays as an enabling technology for attention aware systems , 2006, Comput. Hum. Behav..

[16]  Krzysztof Krejtz,et al.  Using Microsaccades to Estimate Task Difficulty During Visual Search of Layered Surfaces , 2020, IEEE Transactions on Visualization and Computer Graphics.

[17]  Sylvain Lefebvre,et al.  Colored Fused Filament Fabrication , 2017, ACM Trans. Graph..

[18]  Wilson S. Geisler,et al.  Focusing on the essential , 2003, Commun. ACM.

[19]  Hans-Werner Gellersen,et al.  Toward Mobile Eye-Based Human-Computer Interaction , 2010, IEEE Pervasive Computing.

[20]  Erhardt Barth,et al.  A gaze-contingent reading tutor program for children with developmental dyslexia , 2011, APGV '11.

[21]  Marcus A. Magnor,et al.  Gaze-Contingent Computational Displays: Boosting perceptual fidelity , 2016, IEEE Signal Processing Magazine.

[22]  Gary Anthes Augmented reality gets real , 2019, Commun. ACM.

[23]  Kenny Mitchell,et al.  User, metric, and computational evaluation of foveated rendering methods , 2016, SAP.

[24]  Yuanzhen Li,et al.  Measuring visual clutter. , 2007, Journal of vision.

[25]  Lester C. Loschky,et al.  Gaze-Contingent Multiresolutional Displays: An Integrative Review , 2003, Hum. Factors.

[26]  Arzu Çöltekin,et al.  GeoGCD: improved visual search via gaze-contingent display , 2019, ETRA.

[27]  Arzu Çöltekin,et al.  A Testbed Combining Visual Perception Models for Geographic Gaze Contingent Displays , 2015, EuroVis.

[28]  Andrew T. Duchowski,et al.  Gaze-based interaction: A 30 year retrospective , 2018, Comput. Graph..

[29]  Arzu Çöltekin,et al.  An Approach to Modeling Spatial Perception for Geovisualization , 2011 .

[30]  Susanne Boll,et al.  Assisting Drivers with Ambient Take-Over Requests in Highly Automated Driving , 2016, AutomotiveUI.

[31]  David Lindlbauer,et al.  Context-Aware Online Adaptation of Mixed Reality Interfaces , 2019, UIST.

[32]  Francisco Escolano,et al.  How do image complexity, task demands and looking biases influence human gaze behavior? , 2013, Pattern Recognit. Lett..

[33]  Zhou Wang,et al.  Modern Image Quality Assessment , 2006, Modern Image Quality Assessment.

[34]  Andreas Bulling,et al.  Moment-to-Moment Detection of Internal Thought during Video Viewing from Eye Vergence Behavior , 2019, ACM Multimedia.

[35]  Heinrich H. Bülthoff,et al.  Eye movement planning on Single-Sensor-Single-Indicator displays is vulnerable to user anxiety and cognitive load , 2017, Journal of Eye Movement Research.

[36]  Timothy C. Clausner,et al.  Enhancing air traffic displays via perceptual cues , 2008, TAP.

[37]  Martin Raubal,et al.  The Index of Pupillary Activity: Measuring Cognitive Load vis-à-vis Task Difficulty with Pupil Oscillation , 2018, CHI.

[38]  Ashley M. Sherman,et al.  Visual search for arbitrary objects in real scenes , 2011, Attention, perception & psychophysics.

[39]  Lester C. Loschky,et al.  Investigating spatial vision and dynamic attentional selection using a gaze-contingent multiresolutional display. , 2002, Journal of experimental psychology. Applied.

[40]  Arzu Çöltekin,et al.  Foveated gaze-contingent displays for peripheral LOD management, 3D visualization, and stereo imaging , 2007, TOMCCAP.

[41]  Miguel P Eckstein,et al.  Visual search: a retrospective. , 2011, Journal of vision.