Hand orientation redundancy filter applied to hand-shapes dataset

We have created a dataset of frames extracted from videos of Irish Sign Language (ISL) for sign language recognition. The dataset was collected by recording human subjects executing ISL hand-shapes and movements. Frames were extracted from the videos producing a total of 52,688 images for the 23 static common hand-shapes. Given that some of the frames were relativity similar we designed a new method for removing redundant frames based on labelling the hand images by using axis of least inertia - Hand Orientation Redundancy Filter (HORF) - and we compare the results with an iterative method - Iterative Redundancy Filter (IRF). This selection process method selects the most different images in order to keep the dataset diverse. The IRF dataset contains 50,000 images whereas the HORF consists of 27,683 images. Finally, we tested two classifiers over the HORF dataset and compared the results with the IRF dataset.

[1]  Berthold K. P. Horn Robot vision , 1986, MIT electrical engineering and computer science series.

[2]  Lorraine Leeson,et al.  Irish sign language : a cognitive linguistic account , 2012 .

[3]  Noel E. O'Connor,et al.  A Dataset for Irish sign language recognition , 2017 .

[4]  Du-Ming Tsai,et al.  Object recognition by a linear weight classifier , 1995, Pattern Recognit. Lett..

[5]  Andreas Geiger,et al.  Are we ready for autonomous driving? The KITTI vision benchmark suite , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[6]  Mohamed Farouk,et al.  Principal Component Pyramids using Image Blurring for Nonlinearity Reduction in Hand Shape Recognition , 2015 .

[7]  Xiao Liu,et al.  Fine-Grained Video Categorization with Redundancy Reduction Attention , 2018, ECCV.

[8]  Noel E. O'Connor,et al.  Irish Sign Language Recognition Using Principal Component Analysis and Convolutional Neural Networks , 2017, 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA).

[9]  Noel E. O'Connor,et al.  A comparison between end-to-end approaches and feature extraction based approaches for Sign Language recognition , 2017, 2017 International Conference on Image and Vision Computing New Zealand (IVCNZ).