Sign language recognition through kinect based depth images and neural network

Sign language is the language of the people with hearing and speaking disabilities. In it mostly hands are moved in a particular way which along with some facial expression produces a meaningful thought which the speaker would like to convey to others. Using the sign language people with speaking and hearing disabilities can communicate with others who know the language very easily but it becomes difficult when it comes to interacting with a normal person. As a result there is a requirement of an intermediate system which will help in improving the interaction between people with the hearing disabilities as well as with the normal people. In this paper we present a sign language recognition technique using kinect depth camera and neural network. Using the kinect camera we obtain the image of the person standing in front of the camera and then we crop the hand region from the depth image and pre-process that image using the morphological operations to remove unwanted region from the hand image and find the contour of the hand sign and from the particular contour position of the hand we generate a signal on which Discrete Cosine Transform (DCT) is applied and first 200 DCT coefficient of the signal are feed to the neural network for training and classification and finally the network classify and recognize the sign. A data set of sign from 0 to 9 are formed using kinect camera and we tested on 1236 images in the database on which training is applied and we achieved 98% training and an average accuracy for all the sign recognition as 83.5%.

[1]  Nianjun Liu,et al.  Evaluation of HMM training algorithms for letter hand gesture recognition , 2003, Proceedings of the 3rd IEEE International Symposium on Signal Processing and Information Technology (IEEE Cat. No.03EX795).

[2]  William T. Freeman,et al.  Orientation Histograms for Hand Gesture Recognition , 1995 .

[3]  Tanner Bryce Blair,et al.  Innovate engineering outreach: A special application of the Xbox 360 Kinect sensor , 2013, 2013 IEEE Frontiers in Education Conference (FIE).

[4]  Zhenjun Tang,et al.  A Fast Convex Hull Algorithm for Binary Image , 2010, Informatica.

[5]  C. Harshith,et al.  Survey on Various Gesture Recognition Techniques for Interfacing Machines Based on Ambient Intelligence , 2010, ArXiv.

[6]  Paul A. Beardsley,et al.  Computer Vision for Interactive Computer Graphics , 1998, IEEE Computer Graphics and Applications.

[7]  Chin-Chen Chang,et al.  New Approach for Static Gesture Recognition , 2006, J. Inf. Sci. Eng..

[8]  Jitendra Malik,et al.  Shape matching and object recognition using shape contexts , 2010, 2010 3rd International Conference on Computer Science and Information Technology.

[9]  Anant Agarwal,et al.  Sign language recognition using Microsoft Kinect , 2013, 2013 Sixth International Conference on Contemporary Computing (IC3).

[10]  Jan Flusser,et al.  Pattern recognition by affine moment invariants , 1993, Pattern Recognit..

[11]  Kanad K. Biswas,et al.  Gesture recognition using Microsoft Kinect® , 2011, The 5th International Conference on Automation, Robotics and Applications.

[12]  Vaishali S. Kulkarni,et al.  Appearance Based Recognition of American Sign Language Using Gesture Segmentation , 2010 .

[13]  이강우,et al.  초코파이 광고에 대한 광고 거장과의 대화 , 2011 .

[14]  S. Majumder,et al.  Shape, texture and local movement hand gesture features for Indian Sign Language recognition , 2011, 3rd International Conference on Trendz in Information Sciences & Computing (TISC2011).

[15]  Huan-Chao Keh,et al.  Real-time Hand Gesture Recognition by Shape Context Based Matching and Cost Matrix , 2011, J. Networks.

[16]  Sandy Huang,et al.  Interpreting American Sign Language with Kinect , 2011 .