Gaze-Aware Mixed-Reality: Addressing Privacy Issues with Eye Tracking

General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. For more information, or if you believe that this document breaches copyright, please contact the Bond University research repository coordinator.

[1]  James Irvine,et al.  Privacy Implications of Wearable Health Devices , 2014, SIN.

[2]  Robert J. K. Jacob,et al.  Evaluation of eye gaze interaction , 2000, CHI.

[3]  M. Just,et al.  Eye fixations and cognitive processes , 1976, Cognitive Psychology.

[4]  Raimund Dachselt,et al.  Look & touch: gaze-supported target acquisition , 2012, CHI.

[5]  Grant McKenzie,et al.  A geoprivacy manifesto , 2018, Trans. GIS.

[6]  Hirohiko Kaneko,et al.  Relationship between Emotional State and Pupil Diameter Variability under Various Types of Workload Stress , 2009, HCI.

[7]  Zhiwei Zhu,et al.  A Real-Time Human Stress Monitoring System Using Dynamic Bayesian Network , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[8]  Naphtali Rishe,et al.  Measurement of pupil diameter variations as a physiological indicator of the affective state in a computer user. , 2007, Biomedical sciences instrumentation.

[9]  Armando Barreto,et al.  Off-line and On-line Stress Detection Through Processing of the Pupil Diameter Signal , 2013, Annals of Biomedical Engineering.

[10]  Sören Preibusch,et al.  Privacy considerations for a pervasive eye tracking world , 2014, UbiComp Adjunct.

[11]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Fabian Göbel,et al.  Look There! Be Social and Share , 2019, CHI 2019.

[13]  O. Wolf,et al.  The role of eye fixation in memory enhancement under stress – An eye tracking study , 2017, Neurobiology of Learning and Memory.

[14]  Christoph Hölscher,et al.  Virtual Reality Experiments with Physiological Measures. , 2018, Journal of visualized experiments : JoVE.

[15]  Florian Alt,et al.  Behavioural Biometrics in VR: Identifying People from Body Motion and Relations in Virtual Reality , 2019, CHI.

[16]  Martin Raubal,et al.  The Index of Pupillary Activity: Measuring Cognitive Load vis-à-vis Task Difficulty with Pupil Oscillation , 2018, CHI.

[17]  Albrecht Schmidt,et al.  Interacting with the Computer Using Gaze Gestures , 2007, INTERACT.

[18]  Sebastian Tilch,et al.  Survey of optical indoor positioning systems , 2011, 2011 International Conference on Indoor Positioning and Indoor Navigation.

[19]  J. Olgin,et al.  Smartphone-Based Geofencing to Ascertain Hospitalizations , 2017, Circulation. Cardiovascular quality and outcomes.

[20]  Mutsumi Watanabe,et al.  Facial Visual-Infrared Stereo Vision Fusion Measurement as an Alternative for Physiological Measurement , 2014 .

[21]  Martin Raubal,et al.  Measuring Cognitive Load for Map Tasks Through Pupil Diameter , 2016, GIScience.

[22]  Yusuke Sugano,et al.  AggreGaze: Collective Estimation of Audience Attention on Public Displays , 2016, UIST.

[23]  Lars Kulik,et al.  Location privacy and location-aware computing , 2006 .

[24]  Adrian Basarab,et al.  Towards an automatic early stress recognition system for office environments based on multimodal measurements: A review , 2016, J. Biomed. Informatics.

[25]  M. Land,et al.  The Roles of Vision and Eye Movements in the Control of Activities of Daily Living , 1998, Perception.

[26]  Markus Kattenbeck,et al.  Geographic-Aware Augmented Reality for VGI , 2019, Advances in Cartography and GIScience of the ICA.

[27]  Dominik Bucher,et al.  Captcha Your Location Proof - A Novel Method for Passive Location Proofs in Adversarial Environments , 2018, LBS.