Zero-Shot Human Activity Recognition Using Non-Visual Sensors

Due to significant advances in sensor technology, studies towards activity recognition have gained interest and maturity in the last few years. Existing machine learning algorithms have demonstrated promising results by classifying activities whose instances have been already seen during training. Activity recognition methods based on real-life settings should cover a growing number of activities in various domains, whereby a significant part of instances will not be present in the training data set. However, to cover all possible activities in advance is a complex and expensive task. Concretely, we need a method that can extend the learning model to detect unseen activities without prior knowledge regarding sensor readings about those previously unseen activities. In this paper, we introduce an approach to leverage sensor data in discovering new unseen activities which were not present in the training set. We show that sensor readings can lead to promising results for zero-shot learning, whereby the necessary knowledge can be transferred from seen to unseen activities by using semantic similarity. The evaluation conducted on two data sets extracted from the well-known CASAS datasets show that the proposed zero-shot learning approach achieves a high performance in recognizing unseen (i.e., not present in the training dataset) new activities.

[1]  Changsheng Xu,et al.  I Know the Relationships: Zero-Shot Action Recognition via Two-Stream Graph Convolutional Networks and Knowledge Graphs , 2019, AAAI.

[2]  Grey Giddins,et al.  Statistics , 2016, The Journal of hand surgery, European volume.

[3]  Geoffrey E. Hinton,et al.  Zero-shot Learning with Semantic Output Codes , 2009, NIPS.

[4]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[5]  Jiyoun Lim,et al.  Sensor Data Acquisition and Multimodal Sensor Fusion for Human Activity Recognition Using Deep Learning , 2019, Sensors.

[6]  Tao Xiang,et al.  Learning a Deep Embedding Model for Zero-Shot Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Kyandoghere Kyamakya,et al.  Activity Recognition in Sensor Data Streams for Active and Assisted Living Environments , 2018, IEEE Transactions on Circuits and Systems for Video Technology.

[8]  Sung-Bae Cho,et al.  Human activity recognition with smartphone sensors using deep learning neural networks , 2016, Expert Syst. Appl..

[9]  Martin L. Griss,et al.  NuActiv: recognizing unseen new activities using semantic attribute-based learning , 2013, MobiSys '13.

[10]  Yuji Matsumoto,et al.  Ridge Regression, Hubness, and Zero-Shot Learning , 2015, ECML/PKDD.

[11]  Chen Xu,et al.  The SUN Attribute Database: Beyond Categories for Deeper Scene Understanding , 2014, International Journal of Computer Vision.

[12]  Majid Sarrafzadeh,et al.  Smartwatch Based Activity Recognition Using Active Learning , 2017, 2017 IEEE/ACM International Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE).

[13]  Piyush Rai,et al.  A Generative Approach to Zero-Shot and Few-Shot Action Recognition , 2018, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).

[14]  Rainer Stiefelhagen,et al.  Towards a Fair Evaluation of Zero-Shot Action Recognition Using External Data , 2018, ECCV Workshops.

[15]  Zhigang Liu,et al.  The Jigsaw continuous sensing engine for mobile phone applications , 2010, SenSys '10.

[16]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[17]  David M. W. Powers,et al.  Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation , 2011, ArXiv.

[18]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[19]  Gierad Laput,et al.  Synthetic Sensors: Towards General-Purpose Sensing , 2017, CHI.

[20]  Nicholas D. Lane,et al.  From smart to deep: Robust activity recognition on smartwatches using deep learning , 2016, 2016 IEEE International Conference on Pervasive Computing and Communication Workshops (PerCom Workshops).

[21]  Pietro Perona,et al.  The Caltech-UCSD Birds-200-2011 Dataset , 2011 .

[22]  Chunyan Miao,et al.  A Survey of Zero-Shot Learning , 2019, ACM Trans. Intell. Syst. Technol..

[23]  Yutaka Arakawa,et al.  Low-cost and Device-free Activity Recognition System with Energy Harvesting PIR and Door Sensors , 2016, MobiQuitous.

[24]  Luminita Dumitriu,et al.  Unobtrusive Monitoring the Daily Activity Routine of Elderly People Living Alone, with Low-Cost Binary Sensors , 2019, Sensors.

[25]  Bernt Schiele,et al.  A tutorial on human activity recognition using body-worn inertial sensors , 2014, CSUR.

[26]  Mohsen Nabian A Comparative Study on Machine Learning Classification Models for Activity Recognition , 2017 .

[27]  Martin L. Griss,et al.  Towards zero-shot learning for human activity recognition using semantic attribute sequence model , 2013, UbiComp.

[28]  Anish Hemant Narkhede Supervised Learning in Human Activity Recognition based on Multimodal Body Sensing , 2017 .

[29]  Ken-ichi Funahashi,et al.  On the approximate realization of continuous mappings by neural networks , 1989, Neural Networks.

[30]  Yoshua Bengio,et al.  Zero-data Learning of New Tasks , 2008, AAAI.

[31]  Yang Liu,et al.  Transductive Unbiased Embedding for Zero-Shot Learning , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[32]  Vijay Srinivasan,et al.  Non-invasive sensor solutions for activity recognition in smart homes , 2012 .

[33]  Amaury Lendasse,et al.  Minimal Learning Machine: A novel supervised distance-based approach for regression and classification , 2015, Neurocomputing.

[34]  XiangTao,et al.  Transductive Multi-View Zero-Shot Learning , 2015 .

[35]  Daijin Kim,et al.  Robust human activity recognition from depth video using spatiotemporal multi-fused features , 2017, Pattern Recognit..

[36]  Arun Sahayadhas,et al.  Research on Human Activity Identification Based on Image Processing and Artificial Intelligence , 2018, International Journal of Engineering & Technology.

[37]  Kent Larson,et al.  Activity Recognition in the Home Using Simple and Ubiquitous Sensors , 2004, Pervasive.

[38]  Xiaoqing Feng,et al.  Real time activity recognition on streaming sensor data for smart environments , 2016, 2016 International Conference on Progress in Informatics and Computing (PIC).

[39]  Diane J. Cook,et al.  Transfer learning for activity recognition: a survey , 2013, Knowledge and Information Systems.

[40]  Diane J. Cook,et al.  Activity recognition on streaming sensor data , 2014, Pervasive Mob. Comput..

[41]  Sutharshan Rajasegarar,et al.  Non-invasive sensor based automated smoking activity detection , 2017, 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[42]  Gwenn Englebienne,et al.  Learning to Recognize Human Activities from Soft Labeled Data , 2014, Robotics: Science and Systems.

[43]  Ibrahim Alper Dogru,et al.  Human Activity Recognition Using Smartphones , 2018, 2018 2nd International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT).

[44]  Yuto Lim,et al.  A Novel Human Activity Recognition and Prediction in Smart Home Based on Interaction , 2019, Sensors.

[45]  Xiangliang Zhang,et al.  Is Attribute-Based Zero-Shot Learning an Ill-Posed Strategy? , 2016, ECML/PKDD.

[46]  Musa Peker,et al.  Human activity recognition from smart watch sensor data using a hybrid of principal component analysis and random forest algorithm , 2018, Measurement and Control.

[47]  Georgiana Dinu,et al.  Improving zero-shot learning by mitigating the hubness problem , 2014, ICLR.

[48]  Chunyan Miao,et al.  Distribution-Based Semi-Supervised Learning for Activity Recognition , 2019, AAAI.

[49]  Joaquín B. Ordieres Meré,et al.  Non-Invasive Ambient Intelligence in Real Life: Dealing with Noisy Patterns to Help Older People , 2019, Sensors.

[50]  Carmen C. Y. Poon,et al.  Unobtrusive Sensing and Wearable Devices for Health Informatics , 2014, IEEE Transactions on Biomedical Engineering.

[51]  Mohammad Reza Kangavari,et al.  Comprehensive architecture for intelligent adaptive interface in the field of single‐human multiple‐robot interaction , 2018 .

[52]  Ali Farhadi,et al.  Describing objects by their attributes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[53]  Med Salim Bouhlel,et al.  A new hybrid deep learning model for human action recognition , 2020, J. King Saud Univ. Comput. Inf. Sci..

[54]  Hwee Pink Tan,et al.  Deep Activity Recognition Models with Triaxial Accelerometers , 2015, AAAI Workshop: Artificial Intelligence Applied to Assistive Technologies and Smart Environments.

[55]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[56]  Yejin Choi,et al.  Zero-Shot Activity Recognition with Verb Attribute Induction , 2017, EMNLP.

[57]  Kyuchang Kang,et al.  Extensible Hierarchical Method of Detecting Interactive Actions for Video Understanding , 2017 .

[58]  K. H. Walse,et al.  Survey on Soft Computing Approaches for Human Activity Recognition , 2017 .

[59]  Qiang Yang,et al.  Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence Transfer Learning for Activity Recognition via Sensor Mapping , 2022 .

[60]  Bernt Schiele,et al.  Weakly Supervised Recognition of Daily Life Activities with Wearable Sensors , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[61]  Anandhakumar Palanisamy,et al.  Vector space based augmented structural kinematic feature descriptor for human activity recognition in videos , 2018, ETRI Journal.

[62]  Alexandros Nanopoulos,et al.  Hubs in Space: Popular Nearest Neighbors in High-Dimensional Data , 2010, J. Mach. Learn. Res..

[63]  Heinrich C. Mayr,et al.  Behavior Modeling and Reasoning for Ambient Support: HCM-L Modeler , 2014, IEA/AIE.

[64]  Amay J Bandodkar,et al.  Non-invasive wearable electrochemical sensors: a review. , 2014, Trends in biotechnology.

[65]  Zhenghua Chen,et al.  A Novel Semisupervised Deep Learning Method for Human Activity Recognition , 2019, IEEE Transactions on Industrial Informatics.

[66]  Zhigang Liu,et al.  Darwin phones: the evolution of sensing and inference on mobile phones , 2010, MobiSys '10.

[67]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[68]  Yu-Liang Hsu,et al.  Human Daily and Sport Activity Recognition Using a Wearable Inertial Sensor Network , 2018, IEEE Access.

[69]  Agne Paulauskaite-Taraseviciene,et al.  Research on human activity recognition based on image classification methods , 2017 .

[70]  Ling Pei,et al.  Weakly Supervised Human Activity Recognition From Wearable Sensors by Recurrent Attention Learning , 2019, IEEE Sensors Journal.

[71]  Fengming Cao,et al.  Activity Recognition Based on Streaming Sensor Data for Assisted Living in Smart Homes , 2015, 2015 International Conference on Intelligent Environments.

[72]  Christoph H. Lampert,et al.  Learning to detect unseen object classes by between-class attribute transfer , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[73]  Nirmalie Wiratunga,et al.  Zero-Shot Learning with Matching Networks for Open-Ended Human Activity Recognition , 2018, SICSA ReaLX.

[74]  Zoran Obradovic,et al.  A Simple yet Effective Model for Zero-Shot Learning , 2018, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).

[75]  Frederic Ringsleben,et al.  Automated Annotation of Sensor data for Activity Recognition using Deep Learning , 2017, GI-Jahrestagung.

[76]  Narayanan C. Krishnan,et al.  Semantically Aligned Bias Reducing Zero Shot Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).