Bowing-Net: Motion Generation for String Instruments Based on Bowing Information

This paper presents a deep learning based method that generates body motion for string instrument performance from raw audio. In contrast to prior methods which aim to predict joint position from audio, we first estimate information that dictates the bowing dynamics, such as the bow direction and the played string. The final body motion is then determined from this information following a conversion rule. By adopting the bowing information as the target domain, not only is learning the mapping more feasible, but also the produced results have bowing dynamics that are consistent with the given audio. We confirmed that our results are superior to existing methods through extensive experiments.

[1]  Hiroyoshi Miwa,et al.  CG animation for piano performance , 2009, SIGGRAPH '09.

[2]  Li Su,et al.  Temporally Guided Music-to-Body-Movement Generation , 2020, ACM Multimedia.

[3]  Ira Kemelmacher-Shlizerman,et al.  Audio to Body Dynamics , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.