Probabilistic identification of sit-to-stand and stand-to-sit with a wearable sensor

Abstract Identification of human movements is crucial for the design of intelligent devices capable to provide assistance. In this work, a Bayesian formulation, together with a sequential analysis method, is presented for identification of sit-to-stand (SiSt) and stand-to-sit (StSi) activities. This method performs autonomous iterative accumulation of sensor measurements and decision-making processes, while dealing with noise and uncertainty present in sensors. First, the Bayesian formulation is able to identify sit, transition and stand activity states. Second, the transition state, divided into transition phases, is used to identify the state of the human body during SiSt and StSi. These processes employ acceleration signals from an inertial measurement unit attached to the thigh of participants. Validation of our method with experiments in offline, real-time and a simulated environment, shows its capability to identify the human body during SiSt and StSi with an accuracy of 100% and mean response time of 50 ms (5 sensor measurements). In the simulated environment, our approach shows its potential to interact with low-level methods required for robot control. Overall, this work offers a robust framework for intelligent and autonomous systems, capable to recognise the human intent to rise from and sit on a chair, which is essential to provide accurate and fast assistance.

[1]  Marc Bolliger,et al.  Automated stand-up and sit-down detection for robot-assisted body-weight support training with the FLOAT , 2017, 2017 International Conference on Rehabilitation Robotics (ICORR).

[2]  Edward D. Lemaire,et al.  Wearable Mobility Monitoring Using a Multimedia Smartphone Platform , 2011, IEEE Transactions on Instrumentation and Measurement.

[3]  Jian Lu,et al.  Recognizing multi-user activities using wearable sensors in a smart home , 2011, Pervasive Mob. Comput..

[4]  Yasuhisa Hasegawa,et al.  Intention-based walking support for paraplegia patients with Robot Suit HAL , 2007 .

[5]  Tony J. Dodd,et al.  Active sensorimotor control for tactile exploration , 2017, Robotics Auton. Syst..

[6]  Matthew Todd Farrell Pattern classification of terrain during amputee walking , 2013 .

[7]  Xiangrong Shen,et al.  A method for early detection of the initiation of sit-to-stand posture transitions , 2016, Physiological measurement.

[8]  Jorge Dias,et al.  A Bayesian framework for active artificial perception , 2013, IEEE Transactions on Cybernetics.

[9]  Tony J. Dodd,et al.  Active Bayesian perception for angle and position discrimination with a biomimetic fingertip , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[10]  Tony J. Dodd,et al.  Feeling the Shape: Active Exploration Behaviors for Object Recognition With a Robotic Hand , 2018, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[11]  Stuart J. Russell,et al.  Dynamic bayesian networks: representation, inference and learning , 2002 .

[12]  Sebastian Thrun,et al.  Probabilistic robotics , 2002, CACM.

[13]  Antonio Bicchi,et al.  An atlas of physical human-robot interaction , 2008 .

[14]  Edward D. Lemaire,et al.  Improving classification of sit, stand, and lie in a smartphone human activity recognition system , 2015, 2015 IEEE International Symposium on Medical Measurements and Applications (MeMeA) Proceedings.

[15]  Syed Mahfuzul Aziz,et al.  Review of Cyber-Physical System in Healthcare , 2014, Int. J. Distributed Sens. Networks.

[16]  Alfredo Petrosino,et al.  Iris recognition through machine learning techniques: A survey , 2016, Pattern Recognit. Lett..

[17]  Adrian Rubio Solis,et al.  Bayesian perception of touch for control of robot emotion , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[18]  Angelica Munoz-Melendez,et al.  Wearable Inertial Sensors for Human Motion Analysis: A Review , 2016, IEEE Sensors Journal.

[19]  Jeff A. Bilmes,et al.  Recognizing Activities and Spatial Context Using Wearable Sensors , 2006, UAI.

[20]  Wendong Xiao,et al.  Real-time Physical Activity classification and tracking using wearble sensors , 2007, 2007 6th International Conference on Information, Communications & Signal Processing.

[21]  Shyamal Patel,et al.  A review of wearable sensors and systems with application in rehabilitation , 2012, Journal of NeuroEngineering and Rehabilitation.

[22]  Robertw . Mann,et al.  Whole-body movements during rising to standing from sitting. , 1990, Physical therapy.

[23]  K Aminian,et al.  Multi-parametric evaluation of sit-to-stand and stand-to-sit transitions in elderly people. , 2011, Medical engineering & physics.

[24]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[25]  Nicholas P. Fey,et al.  Intent Recognition in a Powered Lower Limb Prosthesis Using Time History Information , 2013, Annals of Biomedical Engineering.

[26]  K Aminian,et al.  Suitability of commercial barometric pressure sensors to distinguish sitting and standing activities for wearable monitoring. , 2014, Medical engineering & physics.

[27]  Hafiz Farhan Maqbool,et al.  A Real-Time Gait Event Detection for Lower Limb Prosthesis Control and Evaluation. , 2016, IEEE transactions on neural systems and rehabilitation engineering : a publication of the IEEE Engineering in Medicine and Biology Society.

[28]  Horst Bunke,et al.  Hidden Markov models: applications in computer vision , 2001 .

[29]  Cheong Boon Soh,et al.  Assessment of Foot Trajectory for Human Gait Phase Detection Using Wireless Ultrasonic Sensor Network , 2016, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[30]  J.K. Aggarwal,et al.  Human activity analysis , 2011, ACM Comput. Surv..

[31]  Robert Riener,et al.  Control strategies for active lower extremity prosthetics and orthotics: a review , 2015, Journal of NeuroEngineering and Rehabilitation.

[32]  Deniz Erdogmus,et al.  The Future of Human-in-the-Loop Cyber-Physical Systems , 2013, Computer.

[33]  M. Munih,et al.  Analysis of standing up and sitting down in humans: definitions and normative data presentation. , 1990, Journal of biomechanics.

[34]  Toshio Fukuda,et al.  Neuro-fuzzy control of a robotic exoskeleton with EMG signals , 2004, IEEE Transactions on Fuzzy Systems.

[35]  D. De Rossi,et al.  Development of a novel algorithm for human fall detection using wearable sensors , 2008, 2008 IEEE Sensors.

[36]  Uriel Martinez-Hernandez,et al.  Simultaneous Bayesian Recognition of Locomotion and Gait Phases With Wearable Sensors , 2018, IEEE Sensors Journal.

[37]  Jian He,et al.  An Unobtrusive Fall Detection and Alerting System Based on Kalman Filter and Bayes Network Classifier , 2017, Sensors.

[38]  Angelo M. Sabatini,et al.  Machine Learning Methods for Classifying Human Physical Activity from On-Body Accelerometers , 2010, Sensors.

[39]  Hiroshi Yokoi,et al.  One-Handed Wearable sEMG Sensor for Myoelectric Control of Prosthetic Hands , 2017 .

[40]  Neil D. Lawrence,et al.  An integrated probabilistic framework for robot perception, learning and memory , 2016, 2016 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[41]  Jack Parker,et al.  The Effectiveness of Lower-Limb Wearable Technology for Improving Activity and Participation in Adult Stroke Survivors: A Systematic Review , 2016, Journal of medical Internet research.

[42]  Zhiquan Wang,et al.  Recognition of human activities using SVM multi-class classifier , 2010, Pattern Recognit. Lett..

[43]  Konrad Paul Kording,et al.  Review TRENDS in Cognitive Sciences Vol.10 No.7 July 2006 Special Issue: Probabilistic models of cognition Bayesian decision theory in sensorimotor control , 2022 .

[44]  Uriel Martinez-Hernandez,et al.  Adaptive Bayesian inference system for recognition of walking activities and prediction of gait events using wearable sensors , 2018, Neural Networks.

[45]  Yasuhisa Hasegawa,et al.  Sit-to-Stand and Stand-to-Sit Transfer Support for Complete Paraplegic Patients with Robot Suit HAL , 2010, Adv. Robotics.

[46]  Uriel Martinez-Hernandez,et al.  Multisensory Wearable Interface for Immersion and Telepresence in Robotics , 2017, IEEE Sensors Journal.

[47]  Mark J. Warshawsky,et al.  A Modern Approach , 2005 .

[48]  Peter Norvig,et al.  Artificial Intelligence: A Modern Approach , 1995 .

[49]  Conor J. Walsh,et al.  Stronger, Smarter, Softer: Next-Generation Wearable Robots , 2014, IEEE Robotics & Automation Magazine.

[50]  Svetha Venkatesh,et al.  Activity recognition and abnormality detection with the switching hidden semi-Markov model , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[51]  Michael Goldfarb,et al.  Multiclass Real-Time Intent Recognition of a Powered Lower Limb Prosthesis , 2010, IEEE Transactions on Biomedical Engineering.

[52]  Rodney A. Brooks,et al.  A Robust Layered Control Syste For A Mobile Robot , 2022 .

[53]  Yeh-Liang Hsu,et al.  Development of a wearable motion detector for telemonitoring and real-time identification of physical activity. , 2009, Telemedicine journal and e-health : the official journal of the American Telemedicine Association.

[54]  James M. Keller,et al.  Sit-to-stand detection using fuzzy clustering techniques , 2010, International Conference on Fuzzy Systems.