Computer Vision for Lifelogging: Characterizing Everyday Activities Based on Visual Semantics

Abstract The rapid development of mobile devices capable of sensing our interactions with the environment has made it possible to assist humans in daily living, for example helping patients with cognitive impairment or providing customized food intake plans for patients with obesity, etc. All of this can be achieved through the passive gathering of detailed records of everyday behavior known as lifelogging. For example, the widely adopted smart mobile phones and newly emerging consumer wearable devices like Google Glass, Baidu Eye, and Narrative Clip are usually embedded with rich sensing capabilities including a camera, accelerometer, GPS, digital compass, etc., which can help to capture daily activity unobtrusively. Among these heterogeneous sensor readings, visual media contain more semantics to assist in characterizing everyday activities, and visual lifelogging is a class of personal sensing that employs wearable cameras to capture images or video sequences of everyday activities. This chapter will focus on the most recent research methods used to understand visual lifelogs, including semantic annotations of visual concepts, use of contextual semantics, recognition of activities, visualization of activities, etc. We also discuss some research challenges that indicate potential directions for future research. This chapter is intended to support readers in the area of assistive living using wearable sensing and computer vision for lifelogging, and human behavior researchers aiming at behavioral analysis based on visual understanding.

[1]  Chong-Wah Ngo,et al.  Fast Semantic Diffusion for Large-Scale Context-Based Image and Video Annotation , 2012, IEEE Transactions on Image Processing.

[2]  Shih-Fu Chang,et al.  A reranking approach for context-based concept fusion in video indexing and retrieval , 2007, CIVR '07.

[3]  Jiang Zhou,et al.  Semantic Indexing of Wearable Camera Images: Kids'Cam Concepts , 2016, iV&L-MM@MM.

[4]  Lifeng Sun,et al.  Characterizing everyday activities from visual lifelogs based on enhancing concept representation , 2016, Comput. Vis. Image Underst..

[5]  Alan F. Smeaton,et al.  Semantics-based selection of everyday concepts in visual lifelogging , 2012, International Journal of Multimedia Information Retrieval.

[6]  Steve Hodges,et al.  Neuropsychological Rehabilitation , 2013 .

[7]  Xiao Li,et al.  Digital Health: Tracking Physiomes and Activity Using Wearable Biosensors Reveals Useful Health-Related Information , 2017, PLoS biology.

[8]  Walter Bender,et al.  Next-Generation Personal Memory Aids , 2004 .

[9]  Paul Kelly,et al.  Developing a Method to Test the Validity of 24 Hour Time Use Diaries Using Wearable Cameras: A Feasibility Pilot , 2015, PloS one.

[10]  Petia Radeva,et al.  LEMoRe: A Lifelog Engine for Moments Retrieval at the NTCIR-Lifelog LSAT Task , 2016, NTCIR.

[11]  Laura A. Dabbish,et al.  Designing games with a purpose , 2008, CACM.

[12]  G. O'loughlin,et al.  Using a wearable camera to increase the accuracy of dietary analysis. , 2013, American journal of preventive medicine.

[13]  Daniela Petrelli,et al.  Socio-Technical Lifelogging: Deriving Design Principles for a Future Proof Digital Past , 2012, Hum. Comput. Interact..

[14]  Alan F. Smeaton,et al.  Constructing a SenseCam visual diary as a media process , 2008, Multimedia Systems.

[15]  S. Marshall,et al.  An ethical framework for automated, wearable cameras in health behavior research. , 2013, American journal of preventive medicine.

[16]  H. S. Bayley,et al.  Four-day multimedia diet records underestimate energy needs in middle-aged and elderly women as determined by doubly-labeled water. , 2000, The Journal of nutrition.

[17]  Noel E. O'Connor,et al.  ShareDay: A Novel Lifelog Management System for Group Sharing , 2013, MMM.

[18]  Jenny Benois-Pineau,et al.  Wearable video monitoring of people with age Dementia : Video indexing at the service of helthcare , 2008, 2008 International Workshop on Content-Based Multimedia Indexing.

[19]  Alan F. Smeaton,et al.  The colour of life: novel visualisations of population lifestyles , 2010, ACM Multimedia.

[20]  David Haussler,et al.  Exploiting Generative Models in Discriminative Classifiers , 1998, NIPS.

[21]  Rami Albatal,et al.  Overview of NTCIR-12 Lifelog Task , 2016, NTCIR.

[22]  H. Sebastian Seung,et al.  Learning the parts of objects by non-negative matrix factorization , 1999, Nature.

[23]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[24]  Ali Farhadi,et al.  You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[25]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[26]  Gang Hua,et al.  Semantic Model Vectors for Complex Video Event Recognition , 2012, IEEE Transactions on Multimedia.

[27]  Georges Quénot,et al.  TRECVid Semantic Indexing of Video: A 6-year Retrospective , 2016 .

[28]  Alan F. Smeaton,et al.  Passively recognising human activities through lifelogging , 2011, Comput. Hum. Behav..

[29]  Chelsea Dobbins,et al.  Exploiting linked data to create rich human digital memories , 2013, Comput. Commun..

[30]  Jacki Liddle,et al.  What older people do: Time use and exploring the link between role participation and life satisfaction in people aged 65 years and over , 2007 .

[31]  Hannah Badland,et al.  Using wearable cameras to categorise type and context of accelerometer-identified episodes of physical activity , 2013, International Journal of Behavioral Nutrition and Physical Activity.

[32]  M. Law,et al.  Occupation, Health and Well-Being , 1998 .

[33]  Yike Guo,et al.  Wiki-health : from quantified self to self-understanding , 2014 .

[34]  E.A.W.H. van den Hoven,et al.  A future-proof past: Designing for remembering experiences: , 2014 .

[35]  Stuart Weibel,et al.  The Dublin Core: A Simple Content Description Model for Electronic Resources , 2005 .

[36]  Steve Hodges,et al.  SenseCam improves memory for recent events and quality of life in a patient with memory retrieval difficulties , 2011, Memory.

[37]  Alan F. Smeaton,et al.  Guidelines for the presentation and visualisation of lifelog content , 2008 .