Attention-driven Expression and Gesture Analysis in an Interactive Environment

To provide natural user interfaces to interactive environments, accurate and fast recognition of gestures and expressions is needed. We adopt a view-based gesture recognition strategy that runs in an un-constrained interactive environment, which uses active vision methods to determine context cuess for the view-based method. Using vision routines already implemented for an interactive environment, we determine the spatial location of salient body parts and guide an active camera to obtain foveated images of gestures or expressions. Face recognition routines used to obtain an estimate of the identity of the user, and provide an index into the best set of view templates to use. The resulting system combines low-resolution, user-independent processing with high-resolution, user-speciic models, all of which are computed in real time as part of an interactive environment.

[1]  F. Girosi,et al.  Networks for approximation and learning , 1990, Proc. IEEE.

[2]  M. Turk,et al.  Eigenfaces for Recognition , 1991, Journal of Cognitive Neuroscience.

[3]  Alex Pentland,et al.  Space-time gestures , 1993, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[4]  Alex Pentland,et al.  View-based and modular eigenspaces for face recognition , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[5]  Alex Pentland,et al.  Correlation and Interpolation Networks for Real-time Expression Analysis/Synthesis , 1994, NIPS.

[6]  Irfan Essa,et al.  Analysis, interpretation and synthesis of facial expressions , 1995 .

[7]  Trevor Darrell,et al.  A novel environment for situated vision and behavior , 1994 .