Learning and Recognition of Human Actions Using Optimal Control Primitives

We propose a unified approach for recognition and learning of human actions, based on an optimal control model of human motion. In this model, the goals and preferences of the agent engaged in a particular action are encapsulated as a cost function or performance criterion, that is optimized to yield the details of the movement. The cost function is a compact, intuitive and flexible representation of the action. A parameterized form of the cost function is considered, wherein the structure reflects the goals of the actions, and the parameters determine the relative weighting of different terms. We show how the cost function parameters can be estimated from data by solving a nonlinear least squares problem. The parameter estimation method is tested on motion capture data for two different reaching actions and six different subjects. We show that the problem of action recognition in the context of this representation is similar to that of mode estimation in a hybrid system and can be solved using a particle filter if a receding horizon formulation of the optimal controller is adopted. We use the proposed approach to recognize different reaching actions from the 3D hand trajectory of subjects.

[1]  Jessica K. Hodgins,et al.  Synthesizing physically realistic human motion in low-dimensional, behavior-specific spaces , 2004, SIGGRAPH 2004.

[2]  Michael I. Jordan,et al.  Optimal feedback control as a theory of motor coordination , 2002, Nature Neuroscience.

[3]  Christoph Bregler,et al.  Learning and recognizing human dynamics in video sequences , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[4]  Andrew Y. Ng,et al.  Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.

[5]  Ruggero Frezza,et al.  A control theory approach to the analysis and synthesis of the experimentally observed motion primitives , 2005, Biological Cybernetics.

[6]  Pieter Abbeel,et al.  Apprenticeship learning via inverse reinforcement learning , 2004, ICML.

[7]  Daniel M. Wolpert,et al.  Signal-dependent noise determines motor planning , 1998, Nature.

[8]  Y. Boers,et al.  Interacting multiple model particle filter , 2003 .

[9]  Richard M. Murray,et al.  A Mathematical Introduction to Robotic Manipulation , 1994 .

[10]  Y. Bar-Shalom,et al.  The interacting multiple model algorithm for systems with Markovian switching coefficients , 1988 .

[11]  Emanuel Todorov,et al.  Iterative Linear Quadratic Regulator Design for Nonlinear Biological Movement Systems , 2004, ICINCO.

[12]  S. Scott Optimal feedback control and the neural basis of volitional motor control , 2004, Nature Reviews Neuroscience.

[13]  G. Rizzolatti,et al.  The mirror-neuron system. , 2004, Annual review of neuroscience.

[14]  N. D. Freitas Rao-Blackwellised particle filtering for fault diagnosis , 2002 .

[15]  M G Pandy,et al.  Optimal control of non-ballistic muscular movements: a constraint-based performance criterion for rising from a chair. , 1995, Journal of biomechanical engineering.

[16]  N. Bergman,et al.  Auxiliary particle filters for tracking a maneuvering target , 2000, Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187).

[17]  Eric Horvitz,et al.  Layered representations for learning and inferring office activity from multiple sensory channels , 2004, Comput. Vis. Image Underst..

[18]  Robert F. Stengel,et al.  Optimal Control and Estimation , 1994 .

[19]  Maja J. Mataric,et al.  Automated Derivation of Primitives for Movement Classification , 2000, Auton. Robots.

[20]  M. Kawato,et al.  Formation and control of optimal trajectory in human multijoint arm movement , 1989, Biological Cybernetics.

[21]  Eyal Amir,et al.  Bayesian Inverse Reinforcement Learning , 2007, IJCAI.

[22]  M. Pandy,et al.  A Dynamic Optimization Solution for Vertical Jumping in Three Dimensions. , 1999, Computer methods in biomechanics and biomedical engineering.

[23]  E. Todorov Optimality principles in sensorimotor control , 2004, Nature Neuroscience.

[24]  M. Pandy,et al.  Dynamic optimization of human walking. , 2001, Journal of biomechanical engineering.

[25]  N. Manning,et al.  The human arm kinematics and dynamics during daily activities - toward a 7 DOF upper limb powered exoskeleton , 2005, ICAR '05. Proceedings., 12th International Conference on Advanced Robotics, 2005..

[26]  M. Pitt,et al.  Filtering via Simulation: Auxiliary Particle Filters , 1999 .

[27]  Aude Billard,et al.  On Learning, Representing, and Generalizing a Task in a Humanoid Robot , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[28]  G. Rizzolatti,et al.  Parietal Lobe: From Action Organization to Intention Understanding , 2005, Science.

[29]  Ruzena Bajcsy,et al.  Recognition of Human Actions using an Optimal Control Based Motor Model , 2008, 2008 IEEE Workshop on Applications of Computer Vision.

[30]  T. Flash,et al.  The coordination of arm movements: an experimentally confirmed mathematical model , 1985, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[31]  Maja J. Mataric,et al.  Exemplar-based primitives for humanoid movement classification and control , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[32]  Rémi Ronfard,et al.  Automatic Discovery of Action Taxonomies from Multiple Views , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[33]  George W. Irwin,et al.  Multiple model bootstrap filter for maneuvering target tracking , 2000, IEEE Trans. Aerosp. Electron. Syst..

[34]  A.D. Kuo,et al.  An optimal control model for analyzing human postural balance , 1995, IEEE Transactions on Biomedical Engineering.

[35]  Anil V. Rao,et al.  Practical Methods for Optimal Control Using Nonlinear Programming , 1987 .