Automatic Detection of Visual Search for the Elderly using Eye and Head Tracking Data

With increasing age we often find ourselves in situations where we search for certain items, such as keys or wallets, but cannot remember where we left them before. Since finding these objects usually results in a lengthy and frustrating process, we propose an approach for the automatic detection of visual search for older adults to identify the point in time when the users need assistance. In order to collect the necessary sensor data for the recognition of visual search, we develop a completely mobile eye and head tracking device specifically tailored to the requirements of older adults. Using this device, we conduct a user study with 30 participants aged between 65 and 80 years ($$avg = 71.7,$$avg=71.7, 50% female) to collect training and test data. During the study, each participant is asked to perform several activities including the visual search for objects in a real-world setting. We use the recorded data to train a support vector machine (SVM) classifier and achieve a recognition rate of 97.55% with the leave-one-user-out evaluation method. The results indicate the feasibility of an approach towards the automatic detection of visual search in the wild.

[1]  A. L. Yarbus Eye Movements During Perception of Complex Objects , 1967 .

[2]  Tobias Baur,et al.  The social signal interpretation (SSI) framework: multimodal signal processing and recognition in real-time , 2013, ACM Multimedia.

[3]  R. C. Langford How People Look at Pictures, A Study of the Psychology of Perception in Art. , 1936 .

[4]  John M. Findlay,et al.  Eye guidance and visual search , 1998 .

[5]  Wei Pan,et al.  SoundSense: scalable sound sensing for people-centric applications on mobile phones , 2009, MobiSys '09.

[6]  Tae-Seong Kim,et al.  A Triaxial Accelerometer-Based Physical-Activity Recognition via Augmented-Signal Features and a Hierarchical Recognizer , 2010, IEEE Transactions on Information Technology in Biomedicine.

[7]  S.Y. Lee,et al.  Accelerometer's position free human activity recognition using a hierarchical recognition model , 2010, The 12th IEEE International Conference on e-Health Networking, Applications and Services.

[8]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[9]  J. Findlay,et al.  Eye guidance and visual search , 1998 .

[10]  Markus Kächele,et al.  Bio-Visual Fusion for Person-Independent Recognition of Pain Intensity , 2015, MCS.

[11]  Johannes Wagner,et al.  Age and gender classification from speech using decision level fusion and ensemble based techniques , 2010, INTERSPEECH.

[12]  Moreno I. Coco,et al.  Classification of visual and linguistic tasks using eye-movement features. , 2014, Journal of vision.

[13]  Johannes Wagner,et al.  A systematic discussion of fusion techniques for multi-modal affect recognition tasks , 2011, ICMI '11.

[14]  Sidney K. D'Mello,et al.  Toward Fully Automated Person-Independent Detection of Mind Wandering , 2014, UMAP.

[15]  J. Beatty Task-evoked pupillary responses, processing load, and the structure of processing resources. , 1982 .

[16]  Yen-Ping Chen,et al.  Online classifier construction algorithm for human activity detection using a tri-axial accelerometer , 2008, Appl. Math. Comput..

[17]  Peter T. Tkacik,et al.  Color Schlieren imaging of high-pressure overexpanded planar nozzle flow using a simple, low-cost test apparatus , 2011, J. Vis..

[18]  Tobias Baur,et al.  Measuring the impact of multimodal behavioural feedback loops on social interactions , 2016, ICMI.

[19]  Udo Seiffert,et al.  A comparison of late fusion methods for object detection , 2013, 2013 IEEE International Conference on Image Processing.

[20]  Anand K. Gramopadhye,et al.  Use of eye movements as feedforward training for a synthetic aircraft inspection task , 2005, CHI.

[21]  Elisabeth André,et al.  Exploring Eye-Tracking-Based Detection of Visual Search for Elderly People , 2016, 2016 12th International Conference on Intelligent Environments (IE).

[22]  Michael D. Dodd,et al.  Examining the influence of task set on eye movements and fixations. , 2011, Journal of Vision.

[23]  Tadahiro Kuroda,et al.  Haar-Like Filtering for Human Activity Recognition Using 3D Accelerometer , 2009, 2009 IEEE 13th Digital Signal Processing Workshop and 5th IEEE Signal Processing Education Workshop.

[24]  Gerhard Tröster,et al.  Eye Movement Analysis for Activity Recognition Using Electrooculography , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Antonio Torralba,et al.  Contextual guidance of eye movements and attention in real-world scenes: the role of global features in object search. , 2006, Psychological review.

[26]  Billur Barshan,et al.  Human Activity Recognition Using Inertial/Magnetic Sensor Units , 2010, HBU.

[27]  Ling Bao,et al.  Activity Recognition from User-Annotated Acceleration Data , 2004, Pervasive.

[28]  Ionut Damian,et al.  Exploring Eye-Tracking-Driven Sonification for the Visually Impaired , 2016, AH.

[29]  James J. Clark,et al.  A computational model for task inference in visual search. , 2013, Journal of vision.

[30]  Koji Yatani,et al.  BodyScope: a wearable acoustic sensor for activity recognition , 2012, UbiComp.

[31]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[32]  Zhenyu He,et al.  Activity recognition from acceleration data based on discrete consine transform and SVM , 2009, 2009 IEEE International Conference on Systems, Man and Cybernetics.

[33]  Veikko Surakka,et al.  Pupil size variation as an indication of affective processing , 2003, Int. J. Hum. Comput. Stud..

[34]  John M. Henderson,et al.  Predicting Cognitive State from Eye Movements , 2013, PloS one.

[35]  Michael L. Littman,et al.  Activity Recognition from Accelerometer Data , 2005, AAAI.

[36]  Bernt Schiele,et al.  Analyzing features for activity recognition , 2005, sOc-EUSAI '05.

[37]  Andrew R. Webb,et al.  Statistical Pattern Recognition , 1999 .

[38]  P. Verghese Visual Search and Attention A Signal Detection Theory Approach , 2001, Neuron.

[39]  Shah Atiqur Rahman,et al.  Unintrusive eating recognition using Google Glass , 2015, 2015 9th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth).

[40]  A. Treisman,et al.  A feature-integration theory of attention , 1980, Cognitive Psychology.

[41]  Ionut Damian,et al.  Social signal processing for dummies , 2016, ICMI.

[42]  Michelle R. Greene,et al.  Reconsidering Yarbus: A failure to predict observers’ task from eye movement patterns , 2012, Vision Research.

[43]  Michael L. Mack,et al.  Viewing task influences eye movement control during active scene perception. , 2009, Journal of vision.

[44]  T. Salthouse,et al.  Decomposing adult age differences in working memory. , 1991 .