An architecture for automatic gesture analysis

The field of human-computer interaction has been widely investigated in the last years, resulting in a variety of systems used in different application fields like virtual reality simulation environments, software user interfaces, and digital library systems. A very crucial part of all these systems is the input module which is devoted to recognize the human operator in terms of tracking and/or recognition of human face, arms position, hand gestures, and so on. In this work a software architecture is presented, for the automatic recognition of human arms poses. Our research has been carried on in the robotics framework. A mobile robot that has to find its path to the goal in a partially structured environment can be trained by a human operator to follow particular routes in order to perform its task quickly. The system is able to recognize and classify some different poses of the operator's arms as direction commands like “turn-left”, “turn-right”, “go-straight”, and so on. A binary image of the operator silhouette is obtained from the gray-level input. Next, a slice centered on the silhouette itself is processed in order to compute the eigenvalues vector of the pixels co-variance matrix. This kind of information is strictly related to the shape of the contour of the operator figure, and can be usefully employed in order to assess the arms' position. Finally, a support vector machine (SVM) is trained in order to classify different poses, using the eigenvalues array. A detailed description of the system is presented along with some remarks on the statistical analysis we used, and on SVM. The experimental results, and an outline of the usability of the system as a generic shape classification tool are also reported.

[1]  Pietro Perona,et al.  Monocular tracking of the human arm in 3D , 1995, Proceedings of IEEE International Conference on Computer Vision.

[2]  David C. Gibbon,et al.  Multi-modal system for locating heads and faces , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[3]  Salvatore Gaglio,et al.  A Cognitive Architecture for Artificial Vision , 1997, Artif. Intell..

[4]  Vladimir Pavlovic,et al.  Visual Interpretation of Hand Gestures for Human-Computer Interaction: A Review , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Alex Pentland,et al.  Real-time self-calibrating stereo person tracking using 3-D shape estimation from blob features , 1996, Proceedings of 13th International Conference on Pattern Recognition.

[6]  Alessandro Verri,et al.  Representing and recognizing visual dynamic events with support vector machines , 1999, Proceedings 10th International Conference on Image Analysis and Processing.

[7]  Franc Solina,et al.  Automatic reconstruction of 3D human arm motion from a monocular image sequence , 1998, Machine Vision and Applications.

[8]  Salvatore Gaglio,et al.  An architecture for autonomous agents exploiting conceptual representations , 1998, Robotics Auton. Syst..

[9]  Jakub Segen,et al.  Gesture based 3D man-machine interaction using a single camera , 1999, Proceedings IEEE International Conference on Multimedia Computing and Systems.

[10]  Fumio Kishino,et al.  Human posture estimation from multiple images using genetic algorithm , 1994, Proceedings of 12th International Conference on Pattern Recognition.

[11]  Bernhard Schölkopf,et al.  Support vector learning , 1997 .

[12]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .