Single and two-person(s) pose estimation based on R-WAA

Human pose estimation methods have difficulties predicting the correct pose for persons due to challenges in scale variation. Existing works in this domain mainly focus on single-person pose estimation. To counter this challenge we have developed a system that can efficiently estimate both one and two individual poses. We termed remarkable joint based, Waveform, Angle, and Alpha characteristics, as R-WAA. R-WAA is a novel up-bottom human pose estimation method developed using two-dimensional body skeletal joint points. They are capturing all required spatial information using waveform characteristics, angle characteristics, and alpha characteristics. All pose estimator characteristics are developed using a remarkable joint, which is the origin of all poses. The proposed algorithm is evaluated for one and two individuals databases: KARD- Kinect Activity Recognition Dataset and SBU Kinect Interaction Dataset. The results of experiments validate that R-WAA outperforms state-of-the-art approaches.

[1]  Alexandre Alahi,et al.  PifPaf: Composite Fields for Human Pose Estimation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  D. Proffitt,et al.  Understanding natural dynamics. , 1989, Journal of experimental psychology. Human perception and performance.

[3]  Xiaohui Xie,et al.  Co-Occurrence Feature Learning for Skeleton Based Action Recognition Using Regularized Deep LSTM Networks , 2016, AAAI.

[4]  Yongqiang Li,et al.  Multiple stream deep learning model for human action recognition , 2020, Image Vis. Comput..

[5]  Hans-Peter Seidel,et al.  VNect , 2017 .

[6]  Björn E. Ottersten,et al.  Localized Trajectories for 2D and 3D Action Recognition † , 2019, Sensors.

[7]  Mohammed Bennamoun,et al.  SkeletonNet: Mining Deep Part Features for 3-D Action Recognition , 2017, IEEE Signal Processing Letters.

[8]  Mohammed Bennamoun,et al.  A New Representation of Skeleton Sequences for 3D Action Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  V. Rao Vemuri,et al.  Use of K-Nearest Neighbor classifier for intrusion detection , 2002, Comput. Secur..

[10]  Jonathan Tompson,et al.  PersonLab: Person Pose Estimation and Instance Segmentation with a Bottom-Up, Part-Based, Geometric Embedding Model , 2018, ECCV.

[11]  Albert A. Rizzo,et al.  FAAST: The Flexible Action and Articulated Skeleton Toolkit , 2011, 2011 IEEE Virtual Reality Conference.

[12]  Chao Li,et al.  Co-occurrence Feature Learning from Skeleton Data for Action Recognition and Detection with Hierarchical Aggregation , 2018, IJCAI.

[13]  M. Shujah Islam Sameem,et al.  CAD: concatenated action descriptor for one and two person(s), using silhouette and silhouette's skeleton , 2020, IET Image Process..

[14]  Jianping Gou,et al.  A new distance-weighted k-nearest neighbor classifier , 2012 .

[15]  Shaohua Wang,et al.  Human action recognition based on scene semantics , 2018, Multimedia Tools and Applications.

[16]  HigherHRNet: Scale-Aware Representation Learning for Bottom-Up Human Pose Estimation , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Zhongfu Ye,et al.  Action recognition using interrelationships of 3D joints and frames based on angle sine relation and distance features using interrelationships , 2021, Applied Intelligence.

[18]  James M. Keller,et al.  A fuzzy K-nearest neighbor algorithm , 1985, IEEE Transactions on Systems, Man, and Cybernetics.

[19]  Marco Morana,et al.  Human Activity Recognition Process Using 3-D Posture Data , 2015, IEEE Transactions on Human-Machine Systems.

[20]  Jing Xu,et al.  Dynamic weighted discrimination power analysis in DCT domain for face and palmprint recognition , 2010, 2010 International Conference on Information and Communication Technology Convergence (ICTC).

[21]  Research on Human Interaction Recognition Algorithm Based on Interest Point of Depth Information Fusion , 2020 .

[22]  Yibin Li,et al.  Region-sequence based six-stream CNN features for general and fine-grained human action recognition in videos , 2018, Pattern Recognit..

[23]  Jonathan Tompson,et al.  Towards Accurate Multi-person Pose Estimation in the Wild , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Jake K. Aggarwal,et al.  Multitype Activity Recognition in Robot-Centric Scenarios , 2015, IEEE Robotics and Automation Letters.

[25]  Saleh Aly,et al.  Human action recognition using bag of global and local Zernike moment features , 2019, Multimedia Tools and Applications.

[26]  Seok-Lyong Lee,et al.  Semantic Image Networks for Human Action Recognition , 2019, International Journal of Computer Vision.

[27]  Gang Yu,et al.  Cascaded Pyramid Network for Multi-person Pose Estimation , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[28]  Luc Van Gool,et al.  Deep Learning on Lie Groups for Skeleton-Based Action Recognition , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  Alberto Del Bimbo,et al.  Submitted to Ieee Transactions on Cybernetics 1 3d Human Action Recognition by Shape Analysis of Motion Trajectories on Riemannian Manifold , 2022 .

[30]  Kaiming He,et al.  Mask R-CNN , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[31]  Gang Wang,et al.  Global Context-Aware Attention LSTM Networks for 3D Action Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  Dong Liu,et al.  Deep High-Resolution Representation Learning for Human Pose Estimation , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[33]  Hailin Guo,et al.  Fall Detection Based on Key Points of Human-Skeleton Using OpenPose , 2020, Symmetry.

[34]  K Ashwini,et al.  Skeletal Data based Activity Recognition System , 2020, 2020 International Conference on Communication and Signal Processing (ICCSP).

[35]  Christian Wolf,et al.  Pose-conditioned Spatio-Temporal Attention for Human Action Recognition , 2017, ArXiv.

[36]  Yichen Wei,et al.  Simple Baselines for Human Pose Estimation and Tracking , 2018, ECCV.

[37]  Zhiao Huang,et al.  Associative Embedding: End-to-End Learning for Joint Detection and Grouping , 2016, NIPS.

[38]  Leif E. Peterson K-nearest neighbor , 2009, Scholarpedia.

[39]  T. R. Garrett,et al.  Normal range of motion of the cervical spine: an initial goniometric study. , 1992, Physical therapy.

[40]  Yaser Sheikh,et al.  OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[41]  Alex Kendall,et al.  What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.

[42]  Yichen Wei,et al.  Integral Human Pose Regression , 2017, ECCV.

[43]  Ennio Gambi,et al.  A Human Activity Recognition System Using Skeleton Data from RGBD Sensors , 2016, Comput. Intell. Neurosci..

[44]  Yong Du,et al.  Skeleton based action recognition with convolutional neural network , 2015, 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR).

[45]  Dimitris Samaras,et al.  Two-person interaction detection using body-pose features and multiple instance learning , 2012, 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.

[46]  Wenjun Zeng,et al.  An End-to-End Spatio-Temporal Attention Model for Human Action Recognition from Skeleton Data , 2016, AAAI.

[47]  Björn E. Ottersten,et al.  Enhanced trajectory-based action recognition using human pose , 2017, 2017 IEEE International Conference on Image Processing (ICIP).

[48]  Saiful Islam,et al.  3D human action analysis and recognition through GLAC descriptor on 2D motion and static posture images , 2019, Multimedia Tools and Applications.

[49]  Hong Cheng,et al.  Interactive body part contrast mining for human interaction recognition , 2014, 2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW).

[50]  Kibum Kim,et al.  Automatic Recognition of Human Interaction via Hybrid Descriptors and Maximum Entropy Markov Model Using Depth Sensors , 2020, Entropy.

[51]  Hans-Peter Seidel,et al.  VNect , 2017, ACM Trans. Graph..

[52]  Norman H. Villaroman,et al.  Teaching natural user interaction using OpenNI and the Microsoft Kinect sensor , 2011, SIGITE '11.