Realtime style transfer for unlabeled heterogeneous human motion

This paper presents a novel solution for realtime generation of stylistic human motion that automatically transforms unlabeled, heterogeneous motion data into new styles. The key idea of our approach is an online learning algorithm that automatically constructs a series of local mixtures of autoregressive models (MAR) to capture the complex relationships between styles of motion. We construct local MAR models on the fly by searching for the closest examples of each input pose in the database. Once the model parameters are estimated from the training data, the model adapts the current pose with simple linear transformations. In addition, we introduce an efficient local regression model to predict the timings of synthesized poses in the output style. We demonstrate the power of our approach by transferring stylistic human motion for a wide variety of actions, including walking, running, punching, kicking, jumping and transitions between those behaviors. Our method achieves superior performance in a comparison against alternative methods. We have also performed experiments to evaluate the generalization ability of our data-driven model as well as the key components of our system.

[1]  Lucas Kovar,et al.  Automated extraction and parameterization of motions in large data sets , 2004, ACM Trans. Graph..

[2]  Jessica K. Hodgins,et al.  Constraint-based motion optimization using a statistical dynamic model , 2007, SIGGRAPH 2007.

[3]  W. Li,et al.  On a mixture autoregressive model , 2000 .

[4]  Lucas Kovar,et al.  Flexible automatic motion blending with registration curves , 2003, SCA '03.

[5]  David J. Fleet,et al.  Multifactor Gaussian process models for style-content separation , 2007, ICML '07.

[6]  Jinxiang Chai,et al.  Accurate realtime full-body motion capture using a single depth camera , 2012, ACM Trans. Graph..

[7]  Carl E. Rasmussen,et al.  Gaussian Processes for Machine Learning (GPML) Toolbox , 2010, J. Mach. Learn. Res..

[8]  David A. Forsyth,et al.  Generalizing motion edits with Gaussian processes , 2009, ACM Trans. Graph..

[9]  Daniel Thalmann,et al.  Using an Intermediate Skeleton and Inverse Kinematics for Motion Retargeting , 2000, Comput. Graph. Forum.

[10]  Kari Pulli,et al.  Style translation for human motion , 2005, SIGGRAPH 2005.

[11]  Aaron Hertzmann,et al.  Style machines , 2000, SIGGRAPH 2000.

[12]  Kenji Amaya,et al.  Emotion from Motion , 1996, Graphics Interface.

[13]  F. Sebastian Grassia,et al.  Practical Parameterization of Rotations Using the Exponential Map , 1998, J. Graphics, GPU, & Game Tools.

[14]  Lucas Kovar,et al.  Motion graphs , 2002, SIGGRAPH '08.

[15]  Michael F. Cohen,et al.  Verbs and Adverbs: Multidimensional Motion Interpolation , 1998, IEEE Computer Graphics and Applications.

[16]  Jinxiang Chai,et al.  Motion graphs++ , 2012, ACM Trans. Graph..

[17]  David W. Aha,et al.  Special Issue on Lazy Learning , 1997 .

[18]  Jinxiang Chai,et al.  Synthesis and editing of personalized stylistic human motion , 2010, I3D '10.

[19]  Jessica K. Hodgins,et al.  Performance animation from low-dimensional control signals , 2005, SIGGRAPH 2005.

[20]  Yong Cao,et al.  Style components , 2006, Graphics Interface.

[21]  Ziv Bar-Joseph,et al.  Modeling spatial and temporal variation in motion data , 2009, ACM Trans. Graph..

[22]  Aaron Hertzmann,et al.  Style-based inverse kinematics , 2004, ACM Trans. Graph..