From beat tracking to beat expectation: Cognitive-based beat tracking for capturing pulse clarity through time

Pulse is the base timing to which western music is commonly notated, generally expressed by a listener by performing periodic taps with their hand or foot. This cognitive construction helps organize the perception of timed events in music and is the most basic expectation in rhythms. The analysis of expectations, and more specifically the strength with which the beat is felt—the pulse clarity—has been used to analyze affect in music. Most computational models of pulse clarity, and rhythmic expectation in general, analyze the input as a whole, without exhibiting changes through a rhythmic passage. We present Tactus Hypothesis Tracker (THT), a model of pulse clarity over time intended for symbolic rhythmic stimuli. The model was developed based on ideas of beat tracking models that extract beat times from musical stimuli. Our model also produces possible beat interpretations for the rhythm, a fitness score for each interpretation and how these evolve in time. We evaluated the model’s pulse clarity by contrasting against tapping variability of human annotators achieving results comparable to a state-of-the-art pulse clarity model. We also analyzed the clarity metric dynamics on synthetic data that introduced changes in the beat, showing that our model presented doubt in the pulse estimation process and adapted accordingly to beat changes. Finally, we assessed if the beat tracking generated by the model was correct regarding listeners tapping data. We compared our beat tracking results with previous beat tracking models. The THT model beat tracking output showed generally correct estimations in phase but exhibits a bias towards a musically correct subdivision of the beat.

[1]  Peter Essens,et al.  Perception of Temporal Patterns , 1985 .

[2]  Aniruddh D. Patel,et al.  The influence of metricality and modality on synchronization with a beat , 2005, Experimental Brain Research.

[3]  David Rosenthal,et al.  Emulation of human rhythm perception , 1992 .

[4]  D. Moelants,et al.  Deviations from the resonance theory of tempo induction , 2004 .

[5]  R. Shepard,et al.  Tonal Schemata in the Perception of Music in Bali and in the West , 1984 .

[6]  Thomas Fillon,et al.  A probabilistic approach to simultaneous extraction of beats and downbeats , 2012, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[7]  D. Temperley The Cognition of Basic Musical Structures , 2001 .

[8]  Godfried T. Toussaint,et al.  Rhythm Complexity Measures: A Comparison of Mathematical Models of Human Perception and Performance , 2008, ISMIR.

[9]  W. Tecumseh Fitch,et al.  Rhythmic cognition in humans and animals: distinguishing meter and pulse perception , 2013, Front. Syst. Neurosci..

[10]  H C Longuet-Higgins,et al.  The Perception of Musical Rhythms , 1982, Perception.

[11]  Simon Dixon,et al.  Evaluation of the Audio Beat Tracking System BeatRoot , 2007 .

[12]  Ernest Mas-Herrero,et al.  Predictability and Uncertainty in the Pleasure of Music: A Reward for Learning? , 2019, The Journal of Neuroscience.

[13]  Luís Paulo Reis,et al.  IBT: A Real-time Tempo and Beat Tracking System , 2010, ISMIR.

[14]  M. Steedman,et al.  The Perception of Musical Rhythm and Metre , 1977, Perception.

[15]  D. Ellis Beat Tracking by Dynamic Programming , 2007 .

[16]  E. Large,et al.  Neural Networks for Beat Perception in Musical Rhythm , 2015, Front. Syst. Neurosci..

[17]  Petri Toiviainen,et al.  A Matlab Toolbox for Music Information Retrieval , 2007, GfKl.

[18]  Ève Poudrier Tapping to Carter: Mensural Determinacy in Complex Rhythmic Sequences , 2018, Empirical Musicology Review.

[19]  R. Zatorre Why Do We Love Music? , 2018, Cerebrum : the Dana forum on brain science.

[20]  Florian Krebs,et al.  madmom: A New Python Audio and Music Signal Processing Library , 2016, ACM Multimedia.

[21]  Patrik Vuilleumier,et al.  Temporal dynamics of musical emotions examined through intersubject synchrony of brain activity. , 2015, Social cognitive and affective neuroscience.

[22]  M. R. Jones,et al.  Time, our lost dimension: toward a new theory of perception, attention, and memory. , 1976, Psychological review.

[23]  Justyna Humięcka-Jakubowska,et al.  Sweet Anticipation : Music and , 2006 .

[24]  Peter A. Martens The Ambiguous Tactus: Tempo, Subdivision Benefit, And Three Listener Strategies , 2011 .

[25]  Vassilis Katsouros,et al.  Convolutional Neural Networks for Real-Time Beat Tracking: A Dancing Robot Application , 2017, ISMIR.

[26]  Florent Meyniel,et al.  The Neural Representation of Sequences: From Transition Probabilities to Algebraic Patterns and Linguistic Trees , 2015, Neuron.

[27]  Florian Krebs,et al.  Joint Beat and Downbeat Tracking with Recurrent Neural Networks , 2016, ISMIR.

[28]  Bruno H Repp,et al.  Rate Limits in Sensorimotor Synchronization With Auditory and Visual Sequences: The Synchronization Threshold and the Benefits and Costs of Interval Subdivision , 2003, Journal of motor behavior.

[29]  Maria A. G. Witek,et al.  Rhythmic complexity and predictive coding: a novel approach to modeling rhythm and meter perception in music , 2014, Front. Psychol..

[30]  Alexander Refsum Jensenius,et al.  Correspondences Between Music and Involuntary Human Micromotion During Standstill , 2018, Front. Psychol..

[31]  Daniel P. W. Ellis,et al.  MIR_EVAL: A Transparent Implementation of Common MIR Metrics , 2014, ISMIR.

[32]  David Margulis,et al.  Musical expectancy and thrills , 2011 .

[33]  J. Sloboda Music Structure and Emotional Response: Some Empirical Findings , 1991 .

[34]  R. Zatorre,et al.  Anatomically distinct dopamine release during anticipation and experience of peak emotion to music , 2011, Nature Neuroscience.

[35]  Henkjan Honing,et al.  Is Beat Induction Innate or Learned? , 2009, Annals of the New York Academy of Sciences.

[36]  Maria A. G. Witek,et al.  Syncopation, Body-Movement and Pleasure in Groove Music , 2014, PloS one.

[37]  Peter Desain,et al.  On tempo tracking: Tempogram Representation and Kalman filtering , 2000, ICMC.

[38]  H. C. Longuet-Higgins,et al.  Perception of melodies , 1976, Nature.

[39]  Leonard B. Meyer Emotion and Meaning in Music , 1957 .

[40]  José Fornari,et al.  Multi-Feature Modeling of Pulse Clarity: Design, Validation and Optimization , 2008, ISMIR.