RapidHARe: A computationally inexpensive method for real-time human activity recognition from wearable sensors

Recent human activity recognition (HAR) methods, based on on-body inertial sensors, have achieved increasing performance; however, this is at the expense of longer CPU calculations and greater energy consumption. Therefore, these complex models might not be suitable for real-time prediction in mobile systems, e.g., in elder-care support and long-term health-monitoring systems. Here, we present a new method called RapidHARe for real-time human activity recognition based on modeling the distribution of a raw data in a half-second context window using dynamic Bayesian networks. Our method does not employ any dynamic-programming-based algorithms, which are notoriously slow for inference, nor does it employ feature extraction or selection methods. In our comparative tests, we show that RapidHARe is an extremely fast predictor, one and a half times faster than artificial neural networks (ANNs) methods, and more than eight times faster than recurrent neural networks (RNNs) and hidden Markov models (HMMs). Moreover, in performance, RapidHare achieves an F1 score of 94.27\% and accuracy of 98.94\%, and when compared to ANN, RNN, HMM, it reduces the F1-score error rate by 45\%, 65\%, and 63\% and the accuracy error rate by 41\%, 55\%, and 62\%, respectively. Therefore, RapidHARe is suitable for real-time recognition in mobile devices.

[1]  Charu C. Aggarwal,et al.  Managing and Mining Sensor Data , 2013, Springer US.

[2]  Vigneshwaran Subbaraju,et al.  Energy-Efficient Continuous Activity Recognition on Mobile Phones: An Activity-Adaptive Approach , 2012, 2012 16th International Symposium on Wearable Computers.

[3]  Ricardo Chavarriaga,et al.  The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition , 2013, Pattern Recognit. Lett..

[4]  G. G. Stokes "J." , 1890, The New Yale Book of Quotations.

[5]  Sinziana Mazilu,et al.  Online detection of freezing of gait with smartphones and machine learning techniques , 2012, 2012 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops.

[6]  Davide Anguita,et al.  Human Activity Recognition on Smartphones Using a Multiclass Hardware-Friendly Support Vector Machine , 2012, IWAAL.

[7]  Angelo M. Sabatini,et al.  Machine Learning Methods for Classifying Human Physical Activity from On-Body Accelerometers , 2010, Sensors.

[8]  Ling Bao,et al.  Activity Recognition from User-Annotated Acceleration Data , 2004, Pervasive.

[9]  Sándor Pongor,et al.  Benchmarking protein classification algorithms via supervised cross-validation. , 2008, Journal of biochemical and biophysical methods.

[10]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[11]  Patrick Olivier,et al.  Slice&Dice: Recognizing Food Preparation Activities Using Embedded Accelerometers , 2009, AmI.

[12]  Jørgen Skotte,et al.  Detection of physical activity types using triaxial accelerometers. , 2014, Journal of physical activity & health.

[13]  Luca Benini,et al.  Activity Recognition from On-Body Sensors: Accuracy-Power Trade-Off by Dynamic Sensor Selection , 2008, EWSN.

[14]  Paul J. M. Havinga,et al.  A Survey of Online Activity Recognition Using Mobile Phones , 2015, Sensors.

[15]  Paul Lukowicz,et al.  Gesture spotting with body-worn inertial sensors to detect user activities , 2008, Pattern Recognit..

[16]  Daniel Roggen,et al.  Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition , 2016, Sensors.

[17]  Zhen Wang,et al.  uWave: Accelerometer-based Personalized Gesture Recognition and Its Applications , 2009, PerCom.

[18]  Tanja Schultz,et al.  Airwriting: a wearable handwriting recognition system , 2013, Personal and Ubiquitous Computing.

[19]  Blake Hannaford,et al.  A Hybrid Discriminative/Generative Approach for Modeling Human Activities , 2005, IJCAI.

[20]  Patrick Olivier,et al.  Feature Learning for Activity Recognition in Ubiquitous Computing , 2011, IJCAI.

[21]  Daniel Olgu ´ õn,et al.  Human Activity Recognition: Accuracy across Common Locations for Wearable Sensors , 2006 .

[22]  Ricardo Chavarriaga,et al.  Benchmarking classification techniques using the Opportunity human activity dataset , 2011, 2011 IEEE International Conference on Systems, Man, and Cybernetics.

[23]  Nicholas Wickström,et al.  A new measure of movement symmetry in early Parkinson's disease patients using symbolic processing of inertial sensor data , 2011, IEEE Transactions on Biomedical Engineering.

[24]  Jeffrey M. Hausdorff,et al.  Wearable Assistant for Parkinson’s Disease Patients With the Freezing of Gait Symptom , 2010, IEEE Transactions on Information Technology in Biomedicine.

[25]  Attila Kertész-Farkas,et al.  HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks , 2017, AIST.

[26]  Andreas Krause,et al.  Trading off prediction accuracy and power consumption for context-aware wearable computing , 2005, Ninth IEEE International Symposium on Wearable Computers (ISWC'05).

[27]  Grégoire Lefebvre,et al.  Inertial Gesture Recognition with BLSTM-RNN , 2015 .

[28]  Anita Sant'Anna,et al.  A Symbolic Approach to Human Motion Analysis Using Inertial Sensors : Framework and Gait Analysis Study , 2012 .

[29]  Gary M. Weiss,et al.  The Impact of Personalization on Smartphone-Based Activity Recognition , 2012, AAAI 2012.

[30]  Susan Coote,et al.  Gait deficits in people with multiple sclerosis: A systematic review and meta-analysis. , 2017, Gait & posture.

[31]  Michael Beigl,et al.  Energy-Efficient Activity Recognition Using Prediction , 2012, 2012 16th International Symposium on Wearable Computers.