Recognition of Real-World Activities from Environmental Sound Cues to Create Life-Log

There are several studies that collect and store life-log for personal memory. This chapter explains about a system that can create someone’s life-log in an inexpensive way to share daily life events with family, friends or care-givers through simple text messaging with a notion to remote monitoring of someone’s wellbeing. In the developed world where people are usually busier than ever, ambient communications through mobile media or the Internet based communication can provide rich social connections to their loving ones ubiquitously whom they care about by sharing awareness information in a passive way. For users who wish to have a persistent existence through ambient communication – to let someone else to know about their daily activity – new technology is needed. Research that aims to simulate virtual living or logging daily events, while challenging and promising, is currently rare. Only very recently the detection of real-world activities has been attempted by processing multiple sensors data along with inference logic for real-world activities. Detecting or inferring human activity using such simple sensor data is often inaccurate, insufficient and expensive. Therefore, this chapter discusses a technology, an inexpensive alternative to other sensors (e.g., accelerometers, proximity sensors etc.) based approaches, to infer human activity from environmental sound cues and common-sense knowledgebase of everyday objects and concepts. A system prototype to log daily events to infer activities in ‘as you go’ manner from environmental sound cues is explained with a few case studies. The input of the system is the patterns of sounds that are usually produced from activities (e.g., toilet flushing), occurring environmentally (e.g., road sounds) or due to interaction with the objects (e.g., cooking utensils clattering). A robust signal processing processes the input sound signal and Hidden Markov Model (HMM) classifiers are developed to detect predetermined sound contexts. Based on the detected sounds and along with the commonsense knowledge regarding human activity, object interaction, ontology of human life (e.g., living pattern of a single old man, or an old couple) and temporal information (e.g., morning, noon etc.) inference engine is employed to detect the activity and the surrounding environment of the person. Preliminary results are encouraging with the accuracy rate for outdoor and indoor related sound categories for activities being above 67% and 61% respectively.

[1]  A. Glascock,et al.  Behavioral Telemedicine: A New Approach to the Continuous Nonintrusive Monitoring of Activities of Daily Living , 2000 .

[2]  E. Campo,et al.  Detecting abnormal behaviour by real-tlme monitoring of patients , 2002 .

[3]  Jeffrey S. Shell,et al.  Augmenting and sharing memory with eyeBlog , 2004, CARPE'04.

[4]  Michael C. Mozer,et al.  The Neural Network House: An Environment that Adapts to its Inhabitants , 1998 .

[5]  Ning Liu,et al.  Bathroom Activity Monitoring Based on Sound , 2005, Pervasive.

[6]  J. Barbenel,et al.  The Use of Artificial Intelligence in the Design of an Intelligent Cognitive Orthosis for People with Dementia , 2001, Assistive technology : the official journal of RESNA.

[7]  Michael A. Cowling,et al.  Non-Speech Environmental Sound Classification System for Autonomous Surveillance , 2004 .

[8]  Alex Pentland,et al.  InSense: Interest-Based Life Logging , 2006, IEEE MultiMedia.

[9]  Jani Mäntyjärvi,et al.  Managing Context Information in Mobile Devices , 2003, IEEE Pervasive Comput..

[10]  Gordon Bell,et al.  Passive capture and ensuing issues for a personal lifetime store , 2004, CARPE'04.

[11]  A. Fitzgibbon,et al.  Memories for life: Managing information over a human lifetime , 2004 .

[12]  Alex Pentland,et al.  Unsupervised clustering of ambulatory audio and video , 1999, 1999 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings. ICASSP99 (Cat. No.99CH36258).

[13]  Gordon Bell,et al.  MyLifeBits: fulfilling the Memex vision , 2002, MULTIMEDIA '02.

[14]  Vannevar Bush,et al.  As we may think , 1945, INTR.

[15]  Biing-Hwang Juang,et al.  Fundamentals of speech recognition , 1993, Prentice Hall signal processing series.

[16]  Karen Zita Haigh,et al.  Learning Models of Human Behaviour with Sequential Patterns , 2002 .

[17]  Guy J. Brown,et al.  Computational Auditory Scene Analysis: Principles, Algorithms, and Applications , 2006 .

[18]  Andrey Temko,et al.  Classification of meeting-room acoustic events with support vector machines and variable-feature-set clustering , 2005, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005..

[19]  Steve Young,et al.  The HTK book , 1995 .

[20]  A. Klapuri,et al.  Audio-based context awareness acoustic modeling and perceptual evaluation , 2003, 2003 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (IEEE Cat. No.03TH8684).

[21]  Sung-Bae Cho,et al.  KeyGraph-based chance discovery for mobile contents management system , 2007, Int. J. Knowl. Based Intell. Eng. Syst..

[22]  Alex Pentland,et al.  The familiar: a living diary and companion , 2001, CHI Extended Abstracts.

[23]  Dadong Wan,et al.  Magic Medicine Cabinet: A Situated Portal for Consumer Healthcare , 1999, HUC.

[24]  Walter Bender,et al.  Next-Generation Personal Memory Aids , 2004 .

[25]  Kathleen R. McKeown,et al.  Predicting the semantic orientation of adjectives , 1997 .

[26]  Thad Starner,et al.  Remembrance Agent: A Continuously Running Automated Information Retrieval System , 1996, PAAM.

[27]  Mika Raento,et al.  ContextPhone: a prototyping platform for context-aware mobile applications , 2005, IEEE Pervasive Computing.

[28]  Gregory Grefenstette,et al.  Validating the Coverage of Lexical Resources for Affect Analysis and Automatically Classifying New Words along Semantic Axes , 2006, Computing Attitude and Affect in Text.

[29]  H.G. Okuno,et al.  Computational Auditory Scene Analysis and its Application to Robot Audition , 2004, 2008 Hands-Free Speech Communication and Microphone Arrays.