Segmenting and Parsing Instrumentalists' Gestures

Abstract This article presents a segmentation model applied to musician movements, taking into account different time structures. In particular we report on ancillary gestures that are not directly linked to sound production, whilst still being entirely part of the global instrumental gesture. Precisely, we study movements of the clarinet captured with an optical 3D motion capture system, analysing ancillary movements assuming that they can be considered as a sequence of primitive actions regarded as base shapes. A stochastic model called the Segmental Hidden Markov Model is used. It allows for the representation of a continuous trajectory as a sequence of primitive temporal profiles taken from a given dictionary. We evaluate the model using two criteria: the Euclidean norm and the log-likelihood, then show that the size of the dictionary does not influence the fitting accuracy and propose a method for building a dictionary based on the log-likelihood criterion. Finally, we show that the sequence of primitive shapes can also be considered as a sequence of symbols enabling us to interpret the data as symbolic patterns and motifs. Based on this representation, we show that circular patterns occur in all players' performances. This symbolic step produces a different layer of interpretation, linked to a larger time scale, which might not be obvious from a direct signal representation.

[1]  Gerhard Widmer,et al.  In Search of the Horowitz Factor , 2003, AI Mag..

[2]  David A. Forsyth,et al.  Motion synthesis from annotations , 2003, ACM Trans. Graph..

[3]  HiltonAdrian,et al.  A survey of advances in vision-based human motion capture and analysis , 2006 .

[4]  Sofia Dahl The Playing of an Accent – Preliminary Observations from Temporal and Kinematic Analysis of Percussionists* , 2000 .

[5]  Michael C. Horsch,et al.  Dynamic Bayesian networks , 1990 .

[6]  A. Gopnik,et al.  Learning from Actions and their Consequences: Inferring Causal Variables from Continuous Sequences of Human Action , 2009 .

[7]  Aaron Hertzmann,et al.  Style machines , 2000, SIGGRAPH 2000.

[8]  Mari Ostendorf,et al.  From HMM's to segment models: a unified view of stochastic modeling for speech recognition , 1996, IEEE Trans. Speech Audio Process..

[9]  Jernej Barbic,et al.  Segmenting Motion Capture Data into Distinct Behaviors , 2004, Graphics Interface.

[10]  Jeffrey M. Zacks,et al.  Segmentation in the perception and memory of events , 2008, Trends in Cognitive Sciences.

[11]  Marcelo M. Wanderley,et al.  The Musical Significance of Clarinetists' Ancillary Gestures: An Exploration of the Field , 2005 .

[12]  Marcelo M. Wanderley,et al.  Gestural control of sound synthesis , 2004, Proceedings of the IEEE.

[13]  Frédéric Bevilacqua,et al.  Effort-based Analysis of Bowing Movements: Evidence of Anticipation Effects , 2008 .

[14]  Caroline Palmer,et al.  Cognitive and biomechanical influences in pianists’ finger tapping , 2007, Experimental Brain Research.

[15]  Norbert Schnell,et al.  Towards a Gesture-Sound Cross-Modal Analysis , 2009, Gesture Workshop.

[16]  Antonio Camurri,et al.  Gesture-Based Communication in Human-Computer Interaction , 2003, Lecture Notes in Computer Science.

[17]  Patrick Gallinari,et al.  Online Handwritten Shape Recognition Using Segmental Hidden Markov Models , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[18]  Taku Komura,et al.  Topology matching for fully automatic similarity estimation of 3D shapes , 2001, SIGGRAPH.

[19]  Frédéric Bevilacqua,et al.  Modeling and segmentation of audio descriptor profiles with segmental models , 2010, Pattern Recognit. Lett..

[20]  Yiannis Aloimonos,et al.  A Language for Human Action , 2007, Computer.

[21]  Eamonn J. Keogh,et al.  Detecting time series motifs under uniform scaling , 2007, KDD '07.

[22]  Ipke Wachsmuth,et al.  Gesture and Sign Language in Human-Computer Interaction , 1998, Lecture Notes in Computer Science.

[23]  Stuart J. Russell,et al.  Dynamic bayesian networks: representation, inference and learning , 2002 .

[24]  Marcelo M. Wanderley,et al.  Trends in Gestural Control of Music , 2000 .

[25]  Rama Chellappa,et al.  Machine Recognition of Human Activities: A Survey , 2008, IEEE Transactions on Circuits and Systems for Video Technology.

[26]  J. F. Soechting,et al.  Anticipatory and sequential motor control in piano playing , 1997, Experimental Brain Research.

[27]  Padhraic Smyth,et al.  Segmental Hidden Markov Models with Random Effects for Waveform Modeling , 2006, J. Mach. Learn. Res..

[28]  J. Davidson Visual Perception of Performance Manner in the Movements of Solo Musicians , 1993 .

[29]  Adrian Hilton,et al.  A survey of advances in vision-based human motion capture and analysis , 2006, Comput. Vis. Image Underst..

[30]  Maja J. Mataric,et al.  Automated Derivation of Primitives for Movement Classification , 2000, Auton. Robots.

[31]  Marcelo M. Wanderley,et al.  Synchronization of motion and timing in clarinet performance , 2009 .

[32]  Alexander Refsum Jensenius,et al.  Musical Gestures: concepts and methods in research , 2010 .

[33]  Esteban Maestre,et al.  Modeling instrumental gestures: an analysis/synthesis framework for violin bowing , 2009 .

[34]  Daniel Thalmann,et al.  PCA-based walking engine using motion capture data , 2004, Proceedings Computer Graphics International, 2004..

[35]  Alexander Refsum Jensenius,et al.  Chunking in Music by Coarticulation , 2010 .

[36]  Marcelo M. Wanderley,et al.  Quantitative Analysis of Non-obvious Performer Gestures , 2001, Gesture Workshop.

[37]  Gerhard Widmer,et al.  Computational Models of Expressive Music Performance: The State of the Art , 2004 .

[38]  Marcelo M. Wanderley,et al.  Performance Gestures of Musicians: What Structural and Emotional Information Do They Convey? , 2003, Gesture Workshop.