Impact of Sliding Window Length in Indoor Human Motion Modes and Pose Pattern Recognition Based on Smartphone Sensors

Human activity recognition (HAR) is essential for understanding people’s habits and behaviors, providing an important data source for precise marketing and research in psychology and sociology. Different approaches have been proposed and applied to HAR. Data segmentation using a sliding window is a basic step during the HAR procedure, wherein the window length directly affects recognition performance. However, the window length is generally randomly selected without systematic study. In this study, we examined the impact of window length on smartphone sensor-based human motion and pose pattern recognition. With data collected from smartphone sensors, we tested a range of window lengths on five popular machine-learning methods: decision tree, support vector machine, K-nearest neighbor, Gaussian naïve Bayesian, and adaptive boosting. From the results, we provide recommendations for choosing the appropriate window length. Results corroborate that the influence of window length on the recognition of motion modes is significant but largely limited to pose pattern recognition. For motion mode recognition, a window length between 2.5–3.5 s can provide an optimal tradeoff between recognition performance and speed. Adaptive boosting outperformed the other methods. For pose pattern recognition, 0.5 s was enough to obtain a satisfactory result. In addition, all of the tested methods performed well.

[1]  Y. Kawahara,et al.  Recognizing User Context Using Mobile Handsets with Acceleration Sensors , 2007, 2007 IEEE International Conference on Portable Information Devices.

[2]  Jeen-Shing Wang,et al.  Activity Recognition Using One Triaxial Accelerometer: A Neuro-fuzzy Classifier with Feature Reduction , 2007, ICEC.

[3]  Naser El-Sheimy,et al.  Towards Arbitrary Placement of Multi-sensors Assisted Mobile Navigation System , 2010 .

[4]  Y.-K. Lee,et al.  Human Activity Recognition via an Accelerometer-Enabled-Smartphone Using Kernel Discriminant Analysis , 2010, 2010 5th International Conference on Future Information Technology.

[5]  J O'gradyMichael,et al.  Dynamic sensor event segmentation for real-time activity recognition in a smart home context , 2015 .

[6]  Jun Zhong,et al.  Towards unsupervised physical activity recognition using smartphone accelerometers , 2016, Multimedia Tools and Applications.

[7]  Yufei Chen,et al.  Performance Analysis of Smartphone-Sensor Behavior for Human Activity Recognition , 2017, IEEE Access.

[8]  M. Stone Asymptotics for and against cross-validation , 1977 .

[9]  Daniel P. Siewiorek,et al.  Activity recognition and monitoring using multiple sensors on different body positions , 2006, International Workshop on Wearable and Implantable Body Sensor Networks (BSN'06).

[10]  María de Lourdes Martínez-Villaseñor,et al.  A Flexible Approach for Human Activity Recognition Using Artificial Hydrocarbon Networks , 2016, Sensors.

[11]  Sylvain Arlot,et al.  A survey of cross-validation procedures for model selection , 2009, 0907.4728.

[12]  Bo Yu,et al.  Convolutional Neural Networks for human activity recognition using mobile sensors , 2014, 6th International Conference on Mobile Computing, Applications and Services.

[13]  William Harwin,et al.  Physical activity recognition of elderly people and people with parkinson's (PwP) during standard mobility tests using wearable sensors , 2017, 2017 International Smart Cities Conference (ISC2).

[14]  Taesoo Lee,et al.  Context Awareness of Human Motion States Using Accelerometer , 2007, Journal of Medical Systems.

[15]  Tahani Daghistani,et al.  Improving Accelerometer-Based Activity Recognition by Using Ensemble of Classifiers , 2016 .

[16]  Seth J. Teller,et al.  Online pose classification and walking speed estimation using handheld devices , 2012, UbiComp '12.

[17]  Majid Sarrafzadeh,et al.  User-centric exergaming with fine-grain activity recognition: a dynamic optimization approach , 2014, UbiComp Adjunct.

[18]  Nasser Kehtarnavaz,et al.  A survey of depth and inertial sensor fusion for human action recognition , 2015, Multimedia Tools and Applications.

[19]  Jacques Georgy,et al.  Estimation of heading misalignment between a pedestrian and a wearable device , 2014, International Conference on Localization and GNSS 2014 (ICL-GNSS 2014).

[20]  Yu Wang,et al.  An Automatic User-Adapted Physical Activity Classification Method Using Smartphones , 2017, IEEE Transactions on Biomedical Engineering.

[21]  Minoru Yoshizawa,et al.  Parameter exploration for response time reduction in accelerometer-based activity recognition , 2013, UbiComp.

[22]  Hongyi Li,et al.  A method to deal with installation errors of wearable accelerometers for human activity recognition , 2011, Physiological measurement.

[23]  Sang Min Yoon,et al.  Human activity recognition from accelerometer data using Convolutional Neural Network , 2017, 2017 IEEE International Conference on Big Data and Smart Computing (BigComp).

[24]  Guang-Zhong Yang,et al.  Sensor Positioning for Activity Recognition Using Wearable Accelerometers , 2011, IEEE Transactions on Biomedical Circuits and Systems.

[25]  James A. Landay,et al.  The Mobile Sensing Platform: An Embedded Activity Recognition System , 2008, IEEE Pervasive Computing.

[26]  Teerayut Horanont,et al.  A classification of accelerometer data to differentiate pedestrian state , 2016, 2016 International Computer Science and Engineering Conference (ICSEC).

[27]  Héctor Pomares,et al.  Window Size Impact in Human Activity Recognition , 2014, Sensors.

[28]  Silvia Lindtner,et al.  Fish'n'Steps: Encouraging Physical Activity with an Interactive Computer Game , 2006, UbiComp.

[29]  Friedrich Foerster,et al.  Detection of posture and motion by accelerometry : a validation study in ambulatory monitoring , 1999 .

[30]  Gary M. Weiss,et al.  Activity recognition using cell phone accelerometers , 2011, SKDD.

[31]  Jake K. Aggarwal,et al.  Human Activity Recognition , 2005, PReMI.

[32]  Bin Liu,et al.  A two-layer and multi-strategy framework for human activity recognition using smartphone , 2016, 2016 IEEE International Conference on Communications (ICC).

[33]  Daijin Kim,et al.  Robust human activity recognition from depth video using spatiotemporal multi-fused features , 2017, Pattern Recognit..

[34]  K. Aminian,et al.  Temporal feature estimation during walking using miniature accelerometers: an analysis of gait improvement after hip arthroplasty , 1999, Medical & Biological Engineering & Computing.

[35]  Katarzyna Radecka,et al.  A Comprehensive Analysis on Wearable Acceleration Sensors in Human Activity Recognition , 2017, Sensors.

[36]  Jian-Jiun Ding,et al.  Real time accelerometer-based gait recognition using adaptive windowed wavelet transforms , 2012, 2012 IEEE Asia Pacific Conference on Circuits and Systems.

[37]  Agathoniki Trigoni,et al.  Robust pedestrian dead reckoning (R-PDR) for arbitrary mobile device placement , 2014, 2014 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[38]  Ophir Frieder,et al.  On understanding and classifying web queries , 2006 .

[39]  Sung-Bae Cho,et al.  Activity Recognition Using Hierarchical Hidden Markov Models on a Smartphone with 3D Accelerometer , 2011, HAIS.

[40]  Tae-Seong Kim,et al.  Accelerometer’s position independent physical activity recognition system for long-term activity monitoring in the elderly , 2010, Medical & Biological Engineering & Computing.

[41]  J. B. J. Bussmann,et al.  Measuring daily behavior using ambulatory accelerometry: The Activity Monitor , 2001, Behavior research methods, instruments, & computers : a journal of the Psychonomic Society, Inc.

[42]  อนิรุธ สืบสิงห์,et al.  Data Mining Practical Machine Learning Tools and Techniques , 2014 .

[43]  T Togawa,et al.  Classification of waist-acceleration signals in a continuous walking record. , 2000, Medical engineering & physics.

[44]  Gregory M. P. O'Hare,et al.  Dynamic sensor event segmentation for real-time activity recognition in a smart home context , 2014, Personal and Ubiquitous Computing.

[45]  Ling Bao,et al.  Activity Recognition from User-Annotated Acceleration Data , 2004, Pervasive.

[46]  Lin Sun,et al.  Activity Recognition on an Accelerometer Embedded Mobile Phone with Varying Positions and Orientations , 2010, UIC.

[47]  David W. McDonald,et al.  Activity sensing in the wild: a field trial of ubifit garden , 2008, CHI.

[48]  H. Hotelling Analysis of a complex of statistical variables into principal components. , 1933 .

[49]  Michael L. Littman,et al.  Activity Recognition from Accelerometer Data , 2005, AAAI.

[50]  Qiang Yang,et al.  Sensor-Based Abnormal Human-Activity Detection , 2008, IEEE Transactions on Knowledge and Data Engineering.

[51]  Juha Röning,et al.  User-Independent Human Activity Recognition Using a Mobile Phone: Offline Recognition vs. Real-Time on Device Recognition , 2012, DCAI.

[52]  Billur Barshan,et al.  Recognizing Daily and Sports Activities in Two Open Source Machine Learning Environments Using Body-Worn Sensor Units , 2014, Comput. J..

[53]  Yan Zhou,et al.  Activity Recognition and Semantic Description for Indoor Mobile Localization , 2017, Sensors.

[54]  Reza Malekian,et al.  Physical Activity Recognition From Smartphone Accelerometer Data for User Context Awareness Sensing , 2017, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[55]  Giancarlo Fortino,et al.  Cloud-based Activity-aaService cyber-physical framework for human activity monitoring in mobility , 2017, Future Gener. Comput. Syst..

[56]  Paul Lukowicz,et al.  Where am I: Recognizing On-body Positions of Wearable Sensors , 2005, LoCA.

[57]  Yen-Ping Chen,et al.  Online classifier construction algorithm for human activity detection using a tri-axial accelerometer , 2008, Appl. Math. Comput..

[58]  Yilmaz Kaya,et al.  Determining the optimal number of body-worn sensors for human activity recognition , 2017, Soft Comput..

[59]  Pietro Siciliano,et al.  An active vision system for fall detection and posture recognition in elderly healthcare , 2010, 2010 Design, Automation & Test in Europe Conference & Exhibition (DATE 2010).

[60]  Wen-Chang Cheng,et al.  Triaxial Accelerometer-Based Fall Detection Method Using a Self-Constructing Cascade-AdaBoost-SVM Classifier , 2013, IEEE Journal of Biomedical and Health Informatics.

[61]  James Brusey,et al.  Postural activity monitoring for increasing safety in bomb disposal missions , 2009 .

[62]  Qiang Shen,et al.  A Handheld Inertial Pedestrian Navigation System With Accurate Step Modes and Device Poses Recognition , 2015, IEEE Sensors Journal.

[63]  Özlem Durmaz Incel,et al.  Multimodal Wireless Sensor Network-Based Ambient Assisted Living in Real Homes with Multiple Residents , 2014, Sensors.

[64]  D. Hand,et al.  Idiot's Bayes—Not So Stupid After All? , 2001 .

[65]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[66]  Aboelmagd Noureldin,et al.  A Survey on Approaches of Motion Mode Recognition Using Sensors , 2017, IEEE Transactions on Intelligent Transportation Systems.

[67]  Hein Putter,et al.  The bootstrap: a tutorial , 2000 .

[68]  Kamiar Aminian,et al.  Spatio-temporal parameters of gait measured by an ambulatory system using miniature gyroscopes. , 2002, Journal of biomechanics.

[69]  Senem Velipasalar,et al.  A Survey on Activity Detection and Classification Using Wearable Sensors , 2017, IEEE Sensors Journal.

[70]  Daniel Roggen,et al.  Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition , 2016, Sensors.

[71]  Daniele Borio,et al.  Accelerometer Signal Features and Classification Algorithms for Positioning Applications , 2011 .

[72]  Sung-Bae Cho,et al.  Human activity recognition with smartphone sensors using deep learning neural networks , 2016, Expert Syst. Appl..

[73]  Meng-Shiuan Pan,et al.  A Step Counting Algorithm for Smartphone Users: Design and Implementation , 2015, IEEE Sensors Journal.

[74]  Korbinian Frank,et al.  Bayesian recognition of safety relevant motion activities with inertial sensors and barometer , 2014, 2014 IEEE/ION Position, Location and Navigation Symposium - PLANS 2014.

[75]  Aboelmagd Noureldin,et al.  Motion Mode Recognition for Indoor Pedestrian Navigation Using Portable Devices , 2016, IEEE Transactions on Instrumentation and Measurement.

[76]  Juan Carlos Niebles,et al.  Sparse composition of body poses and atomic actions for human activity recognition in RGB-D videos , 2017, Image Vis. Comput..

[77]  Jun Yang,et al.  Toward physical activity diary: motion recognition using simple acceleration features with mobile phones , 2009, IMCE '09.