Modelling daily actions through hand-based spatio-temporal features

In this paper, we propose a new approach to domestic action recognition based on a set of features which describe the relation between poses and movements of both hands. These features represent a set of basic actions in a kitchen in terms of the mimics of the hand movements, without needing information of the objects present in the scene. They address specifically the intra-class dissimilarity problem, which occurs when the same action is performed in different ways. The goal is to create a generic methodology that enables a robotic assistant system to recognize actions related to daily life activities and then, be endowed with a proactive behavior. The proposed system uses depth and color data acquired from a Kinect-style sensor and a hand tracking system. We analyze the relevance of the proposed hand-based features using a state-space search approach. Finally, we show the effectiveness of our action recognition approach using our own dataset.

[1]  Dieter Fox,et al.  Fine-grained kitchen activity recognition using RGB-D , 2012, UbiComp.

[2]  Hiroshi Mizoguchi,et al.  Role Analysis of Dominant and Non-dominant Hand in Daily Life , 2013, 2013 IEEE International Conference on Systems, Man, and Cybernetics.

[3]  Mohan M. Trivedi,et al.  The Power Is in Your Hands: 3D Analysis of Hand Gestures in Naturalistic Video , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops.

[4]  Mi Zhang,et al.  A feature selection-based framework for human activity recognition using wearable multimodal sensors , 2011, BODYNETS.

[5]  J. Popp,et al.  Sample size planning for classification models. , 2012, Analytica chimica acta.

[6]  Jake K. Aggarwal,et al.  Human activity recognition from 3D data: A review , 2014, Pattern Recognit. Lett..

[7]  Stephen J. McKenna,et al.  Combining embedded accelerometers with computer vision for recognizing food preparation activities , 2013, UbiComp.

[8]  Mark A. Hall,et al.  Correlation-based Feature Selection for Machine Learning , 2003 .

[9]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[10]  Henry A. Kautz,et al.  Fine-grained activity recognition by aggregating abstract object usage , 2005, Ninth IEEE International Symposium on Wearable Computers (ISWC'05).

[11]  G. Johansson Visual perception of biological motion and a model for its analysis , 1973 .

[12]  Yiannis Aloimonos,et al.  Towards a Watson that sees: Language-guided action recognition for robots , 2012, 2012 IEEE International Conference on Robotics and Automation.

[13]  Gordon Cheng,et al.  Automatic segmentation and recognition of human activities from observation based on semantic reasoning , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[14]  James M. Rehg,et al.  A Scalable Approach to Activity Recognition based on Object Use , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[15]  日向 俊二 Kinect for Windowsアプリを作ろう , 2012 .