BeatBox: end-user interactive definition and training of recognizers for percussive vocalizations

Interactive end-user training of machine learning systems has received significant attention as a tool for personalizing recognizers. However, most research limits end users to training a fixed set of application-defined concepts. This paper considers additional challenges that arise in end-user support for defining the number and nature of concepts that a system must learn to recognize. We develop BeatBox, a new system that enables end-user creation of custom beatbox recognizers and interactive adaptation of recognizers to an end user's technique, environment, and musical goals. BeatBox proposes rapid end-user exploration of variations in the number and nature of learned concepts, and provides end users with feedback on the reliability of recognizers learned for different potential combinations of percussive vocalizations. In a preliminary evaluation, we observed that end users were able to quickly create usable classifiers, that they explored different combinations of concepts to test alternative vocalizations and to refine classifiers for new musical contexts, and that learnability feedback was often helpful in alerting them to potential difficulties with a desired learning concept.

[1]  Perry R. Cook,et al.  Human model evaluation in interactive supervised learning , 2011, CHI.

[2]  Tony DeRose,et al.  Proton: multitouch gestures as regular expressions , 2012, CHI.

[3]  James A. Landay,et al.  Quill: a gesture design tool for pen-based user interfaces , 2001 .

[4]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[5]  Weng-Keen Wong,et al.  Explanatory Debugging: Supporting End-User Debugging of Machine-Learned Programs , 2010, VL/HCC.

[6]  Masataka Goto,et al.  Voice Drummer : A Music Notation Interface of Drum Sounds Using Voice Percussion Input , 2005 .

[7]  Dan Stowell,et al.  Making music through real-time voice timbre analysis: machine learning and timbral control , 2010 .

[8]  Thomas G. Dietterich,et al.  Interacting meaningfully with machine learning systems: Three experiments , 2009, Int. J. Hum. Comput. Stud..

[9]  Ichiro Fujinaga,et al.  Beatbox Classification Using ACE , 2005, ISMIR.

[10]  Desney S. Tan,et al.  Examining multiple potential models in end-user interactive concept learning , 2010, CHI.

[11]  Ajay Kapur,et al.  Query-by-Beat-Boxing: Music Retrieval For The DJ , 2004, ISMIR.

[12]  Desney S. Tan,et al.  CueFlik: interactive concept learning in image search , 2008, CHI.

[13]  Yang Li,et al.  Gesture script: recognizing gestures and their structure using rendering scripts and interactively trained parts , 2014, CHI.

[14]  Thad Starner,et al.  MAGIC: a motion gesture design tool , 2010, CHI.

[15]  Amaury Hazan PERFORMING EXPRESSIVE RHYTHMS WITH BILLABOOP VOICE-DRIVEN DRUM GENERATOR , 2005 .