Enhanced visualisation of dance performance from automatically synchronised multimodal recordings

The Huawei/3DLife Grand Challenge Dataset provides multimodal recordings of Salsa dancing, consisting of audiovisual streams along with depth maps and inertial measurements. In this paper, we propose a system for augmented reality-based evaluations of Salsa dancer performances. An essential step for such a system is the automatic temporal synchronisation of the multiple modalities captured from different sensors, for which we propose efficient solutions. Furthermore, we contribute modules for the automatic analysis of dance performances and present an original software application, specifically designed for the evaluation scenario considered, which enables an enhanced dance visualisation experience, through the augmentation of the original media with the results of our automatic analyses.

[1]  Bernhard Schölkopf,et al.  Learning with kernels , 2001 .

[2]  Takayuki Okatani,et al.  Video Synchronization Based on Co-occurrence of Appearance Changes in Video Sequences , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[3]  Andrew Zisserman,et al.  2D Articulated Human Pose Estimation and Retrieval in (Almost) Unconstrained Still Images , 2012, International Journal of Computer Vision.

[4]  Alexander J. Smola,et al.  Learning with kernels , 1998 .

[5]  Cyril Concolato,et al.  GPAC: open source multimedia framework , 2007, ACM Multimedia.

[6]  Miguel A. Alonso,et al.  Extracting note onsets from musical recordings , 2005, 2005 IEEE International Conference on Multimedia and Expo.

[7]  Steffen Leonhardt,et al.  Automatic Step Detection in the Accelerometer Signal , 2007, BSN.