HF-SPHR: Hybrid Features for Sustainable Physical Healthcare Pattern Recognition Using Deep Belief Networks

The daily life-log routines of elderly individuals are susceptible to numerous complications in their physical healthcare patterns. Some of these complications can cause injuries, followed by extensive and expensive recovery stages. It is important to identify physical healthcare patterns that can describe and convey the exact state of an individual’s physical health while they perform their daily life activities. In this paper, we propose a novel Sustainable Physical Healthcare Pattern Recognition (SPHR) approach using a hybrid features model that is capable of distinguishing multiple physical activities based on a multiple wearable sensors system. Initially, we acquired raw data from well-known datasets, i.e., mobile health and human gait databases comprised of multiple human activities. The proposed strategy includes data pre-processing, hybrid feature detection, and feature-to-feature fusion and reduction, followed by codebook generation and classification, which can recognize sustainable physical healthcare patterns. Feature-to-feature fusion unites the cues from all of the sensors, and Gaussian mixture models are used for the codebook generation. For the classification, we recommend deep belief networks with restricted Boltzmann machines for five hidden layers. Finally, the results are compared with state-of-the-art techniques in order to demonstrate significant improvements in accuracy for physical healthcare pattern recognition. The experiments show that the proposed architecture attained improved accuracy rates for both datasets, and that it represents a significant sustainable physical healthcare pattern recognition (SPHR) approach. The anticipated system has potential for use in human–machine interaction domains such as continuous movement recognition, pattern-based surveillance, mobility assistance, and robot control systems.

[1]  Zhizeng Luo,et al.  Evaluation of Feature Extraction and Recognition for Activity Monitoring and Fall Detection Based on Wearable sEMG Sensors , 2017, Sensors.

[2]  Tae-Seong Kim,et al.  Depth video-based human activity recognition system using translation and scaling invariant features for life logging at smart home , 2012, IEEE Transactions on Consumer Electronics.

[3]  Tae-Seong Kim,et al.  Recognition of Human Home Activities via Depth Silhouettes and ℜ Transformation for Smart Homes , 2012 .

[4]  Kibum Kim,et al.  Automatic Recognition of Human Interaction via Hybrid Descriptors and Maximum Entropy Markov Model Using Depth Sensors , 2020, Entropy.

[5]  Xu Yong,et al.  Three-stage network for age estimation , 2019 .

[6]  Tae-Seong Kim,et al.  Human Activity Recognition via the Features of Labeled Depth Body Parts , 2012, ICOST.

[7]  Xiaohui Peng,et al.  Deep Learning for Sensor-based Activity Recognition: A Survey , 2017, Pattern Recognit. Lett..

[8]  Isaac Debache,et al.  A Lean and Performant Hierarchical Model for Human Activity Recognition Using Body-Mounted Sensors , 2020, Sensors.

[9]  Daijin Kim,et al.  A Depth Video Sensor-Based Life-Logging Human Activity Recognition System for Elderly Care in Smart Indoor Environments , 2014, Sensors.

[10]  Rodrigo Capobianco Guido,et al.  A tutorial review on entropy-based handcrafted feature extraction for information fusion , 2018, Inf. Fusion.

[11]  Kibum Kim,et al.  Automated Sustainable Multi-Object Segmentation and Recognition via Modified Sampling Consensus and Kernel Sliding Perceptron , 2020, Symmetry.

[12]  Hiram Ponce,et al.  A vision-based approach for fall detection using multiple cameras and convolutional neural networks: A case study using the UP-Fall detection dataset , 2019, Comput. Biol. Medicine.

[13]  Ahmad Jalal,et al.  Human Posture Estimation and Sustainable Events Classification via Pseudo-2D Stick Model and K-ary Tree Hashing , 2020, Sustainability.

[14]  Bin Fang,et al.  Gait Neural Network for Human-Exoskeleton Interaction , 2020, Frontiers in Neurorobotics.

[15]  Dwi Esti Kusumandari,et al.  Extraction of P and T Waves from Electrocardiogram Signals with Modified Hamilton Algorithm , 2019, 2019 International Conference on Sustainable Engineering and Creative Computing (ICSECC).

[16]  Attila Kertész-Farkas,et al.  HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks , 2017, AIST.

[17]  Kibum Kim,et al.  A Novel Statistical Method for Scene Classification Based on Multi-Object Categorization and Logistic Regression , 2020, Sensors.

[18]  Duoqian Miao,et al.  Influence of kernel clustering on an RBFN , 2019, CAAI Trans. Intell. Technol..

[19]  Tae-Seong Kim,et al.  Human Activity Recognition via Recognized Body Parts of Human Depth Silhouettes for Residents Monitoring Services at Smart Home , 2013 .

[20]  Kibum Kim,et al.  A Study of Accelerometer and Gyroscope Measurements in Physical Life-Log Activities Detection Systems , 2020, Sensors.

[21]  Héctor Pomares,et al.  mHealthDroid: A Novel Framework for Agile Development of Mobile Health Applications , 2014, IWAAL.

[22]  Mohammad Mehdi Homayounpour,et al.  A brief survey on deep belief networks and introducing a new object oriented MATLAB toolbox (DeeBNet) , 2014, ArXiv.

[23]  L. Minh Dang,et al.  Sensor-based and vision-based human activity recognition: A comprehensive survey , 2020, Pattern Recognit..

[24]  Steven M. Pincus,et al.  A regularity statistic for medical data analysis , 1991, Journal of Clinical Monitoring.

[25]  Hamed Azami,et al.  Refined Composite Multiscale Dispersion Entropy and its Application to Biomedical Signals , 2016, IEEE Transactions on Biomedical Engineering.

[26]  Tan-Hsu Tan,et al.  Device-Free Non-Privacy Invasive Classification of Elderly Travel Patterns in a Smart House Using PIR Sensors and DCNN , 2018, IEEE Sensors Journal.

[27]  Majid Ali Khan Quaid,et al.  Wearable sensors based human behavioral pattern recognition using statistical features and reweighted genetic algorithm , 2019, Multimedia Tools and Applications.

[28]  Stéphane Mallat,et al.  A Theory for Multiresolution Signal Decomposition: The Wavelet Representation , 1989, IEEE Trans. Pattern Anal. Mach. Intell..

[29]  Ahmad Jalal,et al.  Sustainable Wearable System: Human Behavior Modeling for Life-Logging Activities Using K-Ary Tree Hashing Classifier , 2020, Sustainability.

[30]  Xiaodong Yang,et al.  Super Normal Vector for Human Activity Recognition with Depth Cameras , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  Dong-Seong Kim,et al.  Detecting Complex 3D Human Motions with Body Model Low-Rank Representation for Real-Time Smart Activity Monitoring System , 2018, KSII Trans. Internet Inf. Syst..

[32]  Maria Mahmood,et al.  WHITE STAG model: wise human interaction tracking and estimation (WHITE) using spatio-temporal and angular-geometric (STAG) descriptors , 2019, Multimedia Tools and Applications.

[33]  Damith Chinthana Ranasinghe,et al.  Towards deep clustering of human activities from wearables , 2020, SEMWEB.

[34]  Nirmalie Wiratunga,et al.  A knowledge-light approach to personalised and open-ended human activity recognition , 2020, Knowl. Based Syst..

[35]  Hamed Azami,et al.  Dispersion Entropy: A Measure for Time-Series Analysis , 2016, IEEE Signal Processing Letters.

[36]  Kibum Kim,et al.  A Wrist Worn Acceleration Based Human Motion Analysis and Classification for Ambient Smart Home System , 2019, Journal of Electrical Engineering & Technology.

[37]  Shuang Wang,et al.  A Review on Human Activity Recognition Using Vision-Based Method , 2017, Journal of healthcare engineering.

[38]  Jiang Li,et al.  Dimensionality reduction of hyperspectral data using discrete wavelet transform feature extraction , 2002, IEEE Trans. Geosci. Remote. Sens..

[39]  Ahmad Jalal,et al.  Wearable Inertial Sensors for Daily Activity Analysis Based on Adam Optimization and the Maximum Entropy Markov Model , 2020, Entropy.

[40]  Wangxin Yu,et al.  Characterization of Surface EMG Signal Based on Fuzzy Entropy , 2007, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[41]  Ausif Mahmood,et al.  Review of Deep Learning Algorithms and Architectures , 2019, IEEE Access.

[42]  Ahmad Al-Kabbany,et al.  Sensor Type, Axis, and Position-Based Fusion and Feature Selection for Multimodal Human Daily Activity Recognition in Wearable Body Sensor Networks , 2020, Journal of healthcare engineering.

[43]  Daijin Kim,et al.  Robust human activity recognition from depth video using spatiotemporal multi-fused features , 2017, Pattern Recognit..

[44]  Munkhjargal Gochoo,et al.  Modeling Two-Person Segmentation and Locomotion for Stereoscopic Action Identification: A Sustainable Video Surveillance System , 2021, Sustainability.

[45]  Kibum Kim,et al.  Telemonitoring of Daily Activity Using Accelerometer and Gyroscope in Smart Home Environments , 2020, Journal of Electrical Engineering & Technology.

[46]  Jürgen Weber,et al.  Analytical analysis of single-stage pressure relief valves , 2019, International Journal of Hydromechatronics.

[47]  Tobi Delbruck,et al.  Real-time classification and sensor fusion with a spiking deep belief network , 2013, Front. Neurosci..

[48]  Seba Susan,et al.  New shape descriptor in the context of edge continuity , 2019, CAAI Trans. Intell. Technol..

[49]  Nicoletta Saulig,et al.  A novel approach to extracting useful information from noisy TFDs using 2D local entropy measures , 2020, EURASIP Journal on Advances in Signal Processing.

[50]  Zrar Kh. Abdul,et al.  A New Feature Extraction Technique Based on 1D Local Binary Pattern for Gear Fault Detection , 2016 .

[51]  Ying Wah Teh,et al.  Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges , 2018, Expert Syst. Appl..

[52]  Hamed Azami,et al.  Refined multiscale fuzzy entropy based on standard deviation for biomedical signal analysis , 2016, Medical & Biological Engineering & Computing.

[53]  Wing W. Y. Ng,et al.  Neural Network Ensembles for Sensor-Based Human Activity Recognition Within Smart Environments , 2019, Sensors.