Transition activity recognition using fuzzy logic and overlapped sliding window-based convolutional neural networks

Abstract In this paper, we propose a novel approach that can recognize transition activities (e.g., turn to left or right, stand up, and travel down the stairs). Unlike simple activities, the transition activities have unique characteristics that change continuously and occur instantaneously. To recognize the transition activities with these characteristics, we applied convolutional neural network (CNN) that is widely adopted to recognize images, voices, and human activities. In addition, to generate input instances for the CNN model, we developed the overlapped sliding window method, which can accurately recognize the transition activities occurring during a short time. To increase the accuracy of the activity recognition, we have learned CNN models by separating the simple activity and the transition activity. Finally, we adopt fuzzy logic that can be used to handle ambiguous activities. All the procedures of recognizing the elderly’s activities are performed using the data collected by the six sensors embedded in the smartphone. The effectiveness of the proposed approach is shown through experiments. We demonstrate that our approach can improve recognition accuracy of transition activities.

[1]  Marimuthu Palaniswami,et al.  Privacy-Preserving Collaborative Deep Learning with Application to Human Activity Recognition , 2017, CIKM.

[2]  Bernt Schiele,et al.  Analyzing features for activity recognition , 2005, sOc-EUSAI '05.

[3]  Duc A. Tran,et al.  The 11th International Conference on Mobile Systems and Pervasive Computing (MobiSPC-2014) A Study on Human Activity Recognition Using Accelerometer Data from Smartphones , 2014 .

[4]  Shyamanta M. Hazarika,et al.  Efficient extraction of spatial relations for extended objects vis-à-vis human activity recognition in video , 2017, Applied Intelligence.

[5]  Rajiv Kapoor,et al.  Hybrid classifier based human activity recognition using the silhouette and cells , 2015, Expert Syst. Appl..

[6]  Blake Hannaford,et al.  A Hybrid Discriminative/Generative Approach for Modeling Human Activities , 2005, IJCAI.

[7]  Davide Anguita,et al.  Transition-Aware Human Activity Recognition Using Smartphones , 2016, Neurocomputing.

[8]  Thomas G. Dietterich Machine Learning for Sequential Data: A Review , 2002, SSPR/SPR.

[9]  Tae-Seong Kim,et al.  A Triaxial Accelerometer-Based Physical-Activity Recognition via Augmented-Signal Features and a Hierarchical Recognizer , 2010, IEEE Transactions on Information Technology in Biomedicine.

[10]  Min Sheng,et al.  Short-time activity recognition with wearable sensors using convolutional neural network , 2016, VRCAI.

[11]  Christian Wolf,et al.  Sequential Deep Learning for Human Action Recognition , 2011, HBU.

[12]  Edward Sazonov,et al.  Highly Accurate Recognition of Human Postures and Activities Through Classification With Rejection , 2014, IEEE Journal of Biomedical and Health Informatics.

[13]  Lau Bee Theng,et al.  Human activity recognition: A review , 2014, 2014 IEEE International Conference on Control System, Computing and Engineering (ICCSCE 2014).

[14]  Khalil El-Khatib,et al.  A Comparative Analysis of the Impact of Features on Human Activity Recognition with Smartphone Sensors , 2017, WebMedia.

[15]  Juan José Pantrigo,et al.  Convolutional Neural Networks and Long Short-Term Memory for skeleton-based human activity and hand gesture recognition , 2018, Pattern Recognit..

[16]  Arindam Ghosh,et al.  Recognizing Human Activities from Smartphone Sensor Signals , 2014, ACM Multimedia.

[17]  Davide Anguita,et al.  Human Activity Recognition on Smartphones Using a Multiclass Hardware-Friendly Support Vector Machine , 2012, IWAAL.

[18]  Simon Fong,et al.  Improvised methods for tackling big data stream mining challenges: case study of human activity recognition , 2016, The Journal of Supercomputing.

[19]  Donghai Guan,et al.  Activity Recognition Based on Semi-supervised Learning , 2007, 13th IEEE International Conference on Embedded and Real-Time Computing Systems and Applications (RTCSA 2007).

[20]  Sang Min Yoon,et al.  Human activity recognition from accelerometer data using Convolutional Neural Network , 2017, 2017 IEEE International Conference on Big Data and Smart Computing (BigComp).

[21]  Zhaozheng Yin,et al.  Human Activity Recognition Using Wearable Sensors by Deep Convolutional Neural Networks , 2015, ACM Multimedia.

[22]  R Kavitha,et al.  Human Activity Recognition from Sensor data using Random Forest Algorithm , 2017 .

[23]  Andrew W. Fitzgibbon,et al.  Real-time human pose recognition in parts from single depth images , 2011, CVPR 2011.

[24]  William L. Goffe,et al.  SIMANN: FORTRAN module to perform Global Optimization of Statistical Functions with Simulated Annealing , 1992 .

[25]  Diane J. Cook,et al.  Simple and Complex Activity Recognition through Smart Phones , 2012, 2012 Eighth International Conference on Intelligent Environments.

[26]  Zebin Tahmina,et al.  Human activity recognition with inertial sensors using a deep learning approach , 2016 .

[27]  Sung-Bae Cho,et al.  Human activity recognition using smartphone sensors with two-stage continuous hidden Markov models , 2014, 2014 10th International Conference on Natural Computation (ICNC).

[28]  Thomas Plötz,et al.  Deep, Convolutional, and Recurrent Models for Human Activity Recognition Using Wearables , 2016, IJCAI.

[29]  Bo Yu,et al.  Convolutional Neural Networks for human activity recognition using mobile sensors , 2014, 6th International Conference on Mobile Computing, Applications and Services.

[30]  Jaeyoung Yang,et al.  Activity Recognition Based on RFID Object Usage for Smart Mobile Devices , 2011, Journal of Computer Science and Technology.

[31]  Wookey Lee,et al.  Relevance maximization for high-recall retrieval problem: finding all needles in a haystack , 2017, The Journal of Supercomputing.

[32]  Lakshmish Ramaswamy,et al.  A Multi-featured Approach for Wearable Sensor-Based Human Activity Recognition , 2016, 2016 IEEE International Conference on Healthcare Informatics (ICHI).

[33]  Diane J. Cook,et al.  Ambient and smartphone sensor assisted ADL recognition in multi-inhabitant smart environments , 2016, J. Ambient Intell. Humaniz. Comput..

[34]  Tahmina Zebin,et al.  Human activity recognition with inertial sensors using a deep learning approach , 2016, 2016 IEEE SENSORS.

[35]  Rajiv Kapoor,et al.  Recognition of abnormal human activity using the changes in orientation of silhouette in key frames , 2015, 2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom).

[36]  J.K. Aggarwal,et al.  Human activity analysis , 2011, ACM Comput. Surv..

[37]  Manuel P. Cuéllar,et al.  A fuzzy ontology for semantic modelling and recognition of human behaviour , 2014, Knowl. Based Syst..

[38]  Faicel Chamroukhi,et al.  Physical Human Activity Recognition Using Wearable Sensors , 2015, Sensors.