Human Action Recognition Based on Skeleton Information and Multi-Feature Fusion

Action assessment and feedback can effectively assist fitness practitioners in improving exercise benefits. In this paper, we address key challenges in human action recognition and assessment by proposing innovative methods that enhance performance while reducing computational complexity. Firstly, we present Oct-MobileNet, a lightweight backbone network, to overcome the limitations of the traditional OpenPose algorithm’s VGG19 network, which exhibits a large parameter size and high device requirements. Oct-MobileNet employs octave convolution and attention mechanisms to improve the extraction of high-frequency features from the human body contour, resulting in enhanced accuracy with reduced model computational burden. Furthermore, we introduce a novel approach for action recognition that combines skeleton-based information and multiple feature fusion. By extracting spatial geometric and temporal characteristics from actions, we employ a sliding window algorithm to integrate these features. Experimental results show the effectiveness of our approach, demonstrating its ability to accurately recognize and classify various human actions. Additionally, we address the evaluation of traditional fitness exercises, specifically focusing on the BaDunJin movements. We propose a multimodal information-based assessment method that combines pose detection and keypoint analysis. Label sequences are obtained through a pose detector and each frame’s keypoint coordinates are represented as pose vectors. Leveraging multimodal information, including label sequences and pose vectors, we explore action similarity and perform quantitative evaluations to help exercisers assess the quality of their exercise performance.

[1]  Yuliang Zhao,et al.  ANN-Enhanced IoT Wristband for Recognition of Player Identity and Shot Types Based on Basketball Shooting Motion Analysis , 2022, IEEE Sensors Journal.

[2]  Gaoyue Sun Golf Swing Correction Based on Deep Learning Body Posture Recognition , 2021, PRIS.

[3]  Hari Prabhat Gupta,et al.  YogaHelp: Leveraging Motion Sensors for Learning Correct Execution of Yoga With Feedback , 2021, IEEE Transactions on Artificial Intelligence.

[4]  Rui Song,et al.  A fast human action recognition network based on spatio-temporal features , 2020, Neurocomputing.

[5]  Justin Amadeus Albert,et al.  Will You Be My Quarantine: A Computer Vision and Inertial Sensor Based Home Exercise System , 2020, PervasiveHealth.

[6]  Qing Lei,et al.  A Comprehensive Survey of Vision-Based Human Action Recognition Methods , 2019, Sensors.

[7]  Yaser Sheikh,et al.  OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[8]  Xiangyang Xue,et al.  Learning to Score Figure Skating Sport Videos , 2018, IEEE Transactions on Circuits and Systems for Video Technology.

[9]  A. Yeung,et al.  A Review Study on the Beneficial Effects of Baduanjin. , 2017, Journal of alternative and complementary medicine.

[10]  David S. Rosenblum,et al.  From action to activity: Sensor-based activity recognition , 2016, Neurocomputing.

[11]  Derek K. Shaeffer,et al.  MEMS inertial sensors: A tutorial overview , 2013, IEEE Communications Magazine.

[12]  Leif E. Peterson K-nearest neighbor , 2009, Scholarpedia.

[13]  Marti A. Hearst Support vector machines , 1998 .

[14]  K. Esbensen,et al.  Principal component analysis , 1987 .

[15]  Rong Ji,et al.  Research on Basketball Shooting Action Based on Image Feature Extraction and Machine Learning , 2020, IEEE Access.