Information Distribution Within Musical Segments

In research on word recognition, it has been shown that word beginnings have higher information content for word identification than word endings; this asymmetric information distribution within words has been argued to be due to the communicative pressure to allow words in speech to be recognized as early as possible. Through entropy analysis using two representative datasets from Wikifonia and the Essen folksong corpus, we show that musical segments also have higher information content (i.e., higher entropy) in segment beginnings than endings. Nevertheless, this asymmetry was not as dramatic as that found within words, and the highest information content was observed in the middle of the segments (i.e., an inverted U pattern). This effect may be because the first and last notes of a musical segment tend to be tonally stable, with more flexibility in the first note for providing the initial context. The asymmetric information distribution within words has been shown to be an important factor accounting for various asymmetric effects in word reading, such as the left-biased preferred viewing location and optimal viewing position effects. Similarly, the asymmetric information distribution within musical segments is a potential factor that can modulate music reading behavior and should not be overlooked.

[1]  I. Peretz,et al.  The role of contour and intervals in the recognition of melody parts: Evidence from cerebral asymmetries in musicians , 1992, Neuropsychologia.

[2]  R. Bagnold,et al.  The nature and correlation of random distributions , 1983, Proceedings of the Royal Society of London. A. Mathematical and Physical Sciences.

[3]  Emilios Cambouropoulos,et al.  Musical Rhythm: A Formal Model for Determining Local Boundaries, Accents and Metre in a Melodic Surface , 1996, Joint International Conference on Cognitive and Systematic Musicology.

[4]  Joël Pynte,et al.  What determines the eyes’ landing position in words? , 2002, Perception & psychophysics.

[5]  E. Newport,et al.  WORD SEGMENTATION : THE ROLE OF DISTRIBUTIONAL CUES , 1996 .

[6]  Shlomo Dubnov,et al.  Spectral Anticipations , 2006, Computer Music Journal.

[7]  Rodney W. Johnson,et al.  Letter-to-sound rules for automatic translation of english text to phonetics , 1976 .

[8]  Marc Brysbaert,et al.  The Right Visual Field Advantage and the Optimal Viewing Position Effect: On the Relation Between Foveal and Parafoveal Word Recognition , 1996 .

[9]  Erik D. Reichle,et al.  The E-Z Reader model of eye-movement control in reading: Comparisons to other models , 2003, Behavioral and Brain Sciences.

[10]  Paul R. Cohen,et al.  Voting experts: An unsupervised algorithm for segmenting sequences , 2007, Intell. Data Anal..

[11]  Mark D. Plumbley,et al.  Information dynamics: patterns of expectation and surprise in the perception of music , 2009, Connect. Sci..

[12]  K. Rayner Eye Guidance in Reading: Fixation Locations within Words , 1979, Perception.

[13]  Jonathan Grainger,et al.  How Initial Fixation Position Influences Visual Word Recognition: A Comparison of French and Arabic , 1996, Brain and Language.

[14]  J. O'Regan,et al.  Convenient fixation location within isolated words of different length and structure. , 1984, Journal of experimental psychology. Human perception and performance.

[15]  M. Raijmakers Rethinking innateness: A connectionist perspective on development. , 1997 .

[16]  Geraint A. Wiggins,et al.  EXPECTATION IN MELODY: THE INFLUENCE OF CONTEXT AND LEARNING , 2006 .

[17]  C Palmer,et al.  Pitch and temporal contributions to musical phrase perception: Effects of harmony, performance timing, and familiarity , 1987, Perception & psychophysics.

[18]  Frans Wiering,et al.  An Experimental Comparison of Human and Automatic Music Segmentation , 2008 .

[19]  R. Kakigi,et al.  Musical Training Enhances Automatic Encoding of Melodic Contour and Interval Structure , 2004, Journal of Cognitive Neuroscience.

[20]  Ian H. Witten,et al.  Comparing human and computational models of music prediction , 1994 .

[21]  J. Youngblood Style as Information , 1958 .

[22]  C. Krumhansl,et al.  Tracing the dynamic changes in perceived tonal organization in a spatial representation of musical keys. , 1982 .

[23]  E. Narmour The Analysis and Cognition of Melodic Complexity: The Implication-Realization Model , 1992 .

[24]  J. Hsiao Visual field differences in visual word recognition can emerge purely from perceptual learning: Evidence from modeling Chinese character pronunciation , 2011, Brain and Language.

[25]  J. Bartlett,et al.  Recognition of transposed melodies: a key-distance effect in developmental perspective. , 1980, Journal of experimental psychology. Human perception and performance.

[26]  Marc Brys,et al.  Moving beyond Kučera and Francis: A critical evaluation of current word frequency norms and the introduction of a new and improved word frequency measure for American English , 2009 .

[27]  S. Trehub,et al.  Musical context effects in infants and adults: key distance. , 1993, Journal of experimental psychology. Human perception and performance.

[28]  Bret Aarden Dynamic melodic expectancy , 2003 .

[29]  Geraint A. Wiggins,et al.  Melodic Grouping in Music Information Retrieval: New Methods and Applications , 2010, Advances in Music Information Retrieval.

[30]  W. J. Bowling Scale and Contour : Two Components of a Theory of Memory for Melodies , 2005 .

[31]  M. Brent Speech segmentation and word discovery: a computational perspective , 1999, Trends in Cognitive Sciences.

[32]  John Hicks,et al.  Phonological reduction, assimilation, intra-word information structure, and the evolution of the lexicon of English: Why fast speech isn't confusing , 1995 .

[33]  D. Temperley Probabilistic Models of Melodic Interval , 2014 .

[34]  P. J. Hutton,et al.  Speech Synthesis and Recognition Systems , 1987 .

[35]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[36]  T. M. Ellison,et al.  Eye-fixation behavior, lexical storage, and visual word recognition in a split processing model. , 2000, Psychological review.

[37]  Mark L. James,et al.  On the Entropy of Music: An Experiment with Bach Chorale Melodies , 1992 .

[38]  Shlomo Dubnov,et al.  Unified View of Prediction and Repetition Structure in Audio Signals With Application to Interest Point Detection , 2008, IEEE Transactions on Audio, Speech, and Language Processing.

[39]  Martin Rohrmeier,et al.  Implicit Learning and Acquisition of Music , 2012, Top. Cogn. Sci..

[40]  D. Temperley The Cognition of Basic Musical Structures , 2001 .

[41]  S. Trehub,et al.  Key membership and implied harmony in Western tonal music: Developmental perspectives , 1994, Perception & psychophysics.

[42]  Jaime A. Madell,et al.  Eye Movements and Music Reading: Where Do We Look Next? , 2008 .

[43]  J. Bradshaw,et al.  Visual field differences in verbal tasks: Effects of task familiarity and sex of subject , 1978, Brain and Language.

[44]  R. Quirk,et al.  A Corpus of English Conversation , 1980 .

[45]  R. Jackendoff,et al.  A Generative Theory of Tonal Music , 1985 .

[46]  S. Trehub,et al.  A comparison of infants' and adults' sensitivity to western musical structure. , 1992, Journal of experimental psychology. Human perception and performance.

[47]  L. Knopoff,et al.  Entropy as a Measure of Style: The Influence of Sample Length , 1983 .

[48]  L. Rothschild,et al.  The distribution of English dictionary word lengths , 1986 .

[49]  Michael R. Brent,et al.  An Efficient, Probabilistically Sound Algorithm for Segmentation and Word Discovery , 1999, Machine Learning.

[50]  Eugene Narmour,et al.  The Analysis and Cognition of Basic Melodic Structures: The Implication-Realization Model , 1990 .

[51]  M. Pearce,et al.  Sweet Anticipation : Music and the Psychology of Expectation , 2007 .

[52]  Charles P. Bourne,et al.  A Study of the Statistics of Letters in English Words , 1961, Inf. Control..

[53]  R. H. Baayen,et al.  The CELEX Lexical Database (CD-ROM) , 1996 .

[54]  Paul T. von Hippel,et al.  Redefining Pitch Proximity: Tessitura and Mobility as Constraints on Melodic Intervals , 2000 .

[55]  G. Legge,et al.  Mr. Chips: An ideal-observer model of reading , 1997 .

[56]  María Herrojo Ruiz,et al.  Unsupervised statistical learning underpins computational, behavioural, and neural manifestations of musical expectation , 2010, NeuroImage.

[57]  J. Hsiao,et al.  The Modulation of Stimulus Structure on Visual Field Asymmetry Effects: The Case of Chinese Character Recognition , 2013, Quarterly journal of experimental psychology.

[58]  Emmanuel J. Yannakoudakis,et al.  An assessment of N-phoneme statistics in phoneme guessing algorithms which aim to incorporate phonotactic constraints , 1992, Speech Commun..

[59]  Andrew Carstairs-McCarthy An Introduction to English Morphology , 2002 .

[60]  Marc Brysbaert,et al.  Tachistoscopic presentation of verbal stimuli for assessing cerebral dominance: Reliability data and some practical recommendations , 1990, Neuropsychologia.

[61]  J. O'Regan Eye movements and reading. , 1990, Reviews of oculomotor research.

[62]  Yetta Kwailing Wong,et al.  Reading direction is sufficient to account for the optimal viewing position in reading: The case of music reading , 2012, CogSci.

[63]  Ian H. Witten,et al.  Multiple viewpoint systems for music prediction , 1995 .

[64]  Marc Brysbaert,et al.  The Right Visual Field Advantage and the Optimal Viewing Position Effect : On the Relation Between Foveal and Parafoveal Word Recognition , 1996 .

[65]  Petri Toiviainen,et al.  MIDI toolbox : MATLAB tools for music research , 2004 .

[66]  Annabel J. Cohen,et al.  Recognition of Transposed Melodic Sequences , 1976 .

[67]  Marc Brysbaert,et al.  Visual constraints in written word recognition: evidence from the optimal viewing-position effect , 2005 .