Real-Time Hand Pose Recognition Based on a Neural Network Using Microsoft Kinect

The Microsoft Kinect sensor is largely used to detect and recognize body gestures and layout with enough reliability, accuracy and precision in a quite simple way. However, the pretty low resolution of the optical sensors does not allow the device to detect gestures of body parts, such as the fingers of a hand, with the same straightforwardness. Given the clear application of this technology to the field of the user interaction within immersive multimedia environments, there is the actual need to have a reliable and effective method to detect the pose of some body parts. In this paper we propose a method based on a neural network to detect in real time the hand pose, to recognize whether it is closed or not. The neural network is used to process information of color, depth and skeleton coming from the Kinect device. This information is preprocessed to extract some significant feature. The output of the neural network is then filtered with a time average, to reduce the noise due to the fluctuation of the input data. We analyze and discuss three possible implementations of the proposed method, obtaining an accuracy of 90% under good conditions of lighting and background, and even reaching the 95% in best cases, in real time.

[1]  Kourosh Khoshelham,et al.  Accuracy analysis of kinect depth data , 2012 .

[2]  Mark Beale,et al.  Neural Network Toolbox™ User's Guide , 2015 .

[3]  Luc Van Gool,et al.  SURF: Speeded Up Robust Features , 2006, ECCV.

[4]  Marco Morana,et al.  Mobile Interface for Content-Based Image Management , 2010, 2010 International Conference on Complex, Intelligent and Software Intensive Systems.

[5]  Luc Van Gool,et al.  Speeded-Up Robust Features (SURF) , 2008, Comput. Vis. Image Underst..

[6]  Alberto Del Bimbo,et al.  Real-time hand status recognition from RGB-D imagery , 2012, Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012).

[7]  Alessandro Genco,et al.  Interacting with Augmented Environments , 2010, IEEE Pervasive Computing.

[8]  Tasnuva Ahmed,et al.  A Neural Network based Real Time Hand Gesture Recognition System , 2012 .

[9]  Guoqiang Peter Zhang,et al.  Neural networks for classification: a survey , 2000, IEEE Trans. Syst. Man Cybern. Part C.

[10]  Domenico Prattichizzo,et al.  Using Kinect for hand tracking and rendering in wearable haptics , 2011, 2011 IEEE World Haptics Conference.

[11]  Junsong Yuan,et al.  Robust hand gesture recognition based on finger-earth mover's distance with a commodity depth camera , 2011, ACM Multimedia.

[12]  Hélène Paugam-Moisy,et al.  Size of Multilayer Networks for Exact Learning: Analytic Approach , 1996, NIPS.

[13]  Kanad K. Biswas,et al.  Gesture recognition using Microsoft Kinect® , 2011, The 5th International Conference on Automation, Robotics and Applications.

[14]  Yangsheng Xu,et al.  Hand tracking and pose recognition via depth and color information , 2012, 2012 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[15]  Filipe Magalhães,et al.  Hand-Geometry Based Recognition System - A Non Restricted Acquisition Approach , 2012, ICIAR.

[16]  Giuseppe Russo,et al.  A Multichannel Information System to Build and Deliver Rich User-Experiences in Exhibits and Museums , 2011, 2011 International Conference on Broadband and Wireless Computing, Communication and Applications.