Environment capturing with Microsoft Kinect

Constructing virtual scenes that incorporate human and object interaction has traditionally been a time consuming process in computer animation, whereby the motion of an actor is first recorded and any objects used in the scene are then intricately added by an animator. The Microsoft Kinect utilizes a synchronized RGBD stream to provide markerless skeletal tracking of humans, enabling efficient motion capture; however, the problem of capturing environment objects remains unsolved. In this paper, we propose a new framework to segment and track three major types of environment objects using Kinect, namely background planes, stationary objects and dynamic objects. We demonstrate that the motion of an actor and their surrounding environment can be obtained at the same time, saving considerable effort for the animators. Our proposed system is best to be applied to applications involving extensive human-object interactions, such as console games and animation designs.

[1]  C. Karen Liu,et al.  Interactive synthesis of human-object interaction , 2009, SCA '09.

[2]  Andrew W. Fitzgibbon,et al.  Real-time human pose recognition in parts from single depth images , 2011, CVPR 2011.

[3]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[4]  Jean-Claude Latombe,et al.  Planning motions with intentions , 1994, SIGGRAPH.

[5]  Radu Bogdan Rusu,et al.  Semantic 3D Object Maps for Everyday Manipulation in Human Living Environments , 2010, KI - Künstliche Intelligenz.

[6]  Petros Faloutsos,et al.  Interactive motion correction and object manipulation , 2007, SIGGRAPH '08.

[7]  D. Thalmann,et al.  Planning collision-free reaching motions for interactive object manipulation and grasping , 2008, SIGGRAPH '08.

[8]  Larry D. Hostetler,et al.  The estimation of the gradient of a density function, with applications in pattern recognition , 1975, IEEE Trans. Inf. Theory.

[9]  Mark Pauly,et al.  Realtime performance-based facial animation , 2011, ACM Trans. Graph..

[10]  Albert A. Rizzo,et al.  FAAST: The Flexible Action and Articulated Skeleton Toolkit , 2011, 2011 IEEE Virtual Reality Conference.

[11]  Jovan Popovic,et al.  Interactive animation of dynamic manipulation , 2006, SCA '06.

[12]  Craig Gotsman,et al.  Articulated Object Reconstruction and Markerless Motion Capture from Depth Video , 2008, Comput. Graph. Forum.

[13]  Radu Bogdan Rusu,et al.  3D is here: Point Cloud Library (PCL) , 2011, 2011 IEEE International Conference on Robotics and Automation.

[14]  C. J. Taylor,et al.  Segmentation and Analysis of RGB-D data , 2010 .