LJ2M dataset: Toward better understanding of music listening behavior and user mood

Recent years have witnessed a growing interest in modeling user behaviors in multimedia research, emphasizing the need to consider human factors such as preference, activity, and emotion in system development and evaluation. Following this research line, we present in this paper the LiveJournal two-million post (LJ2M) dataset to foster research on user-centered music information retrieval. The new dataset is characterized by the great diversity of real-life listening contexts where people and music interact. It contains blog articles from the social blogging website LiveJournal, along with tags self-reporting a user's emotional state while posting and the musical track that the user considered as the best match for the post. More importantly, the data are contributed by users spontaneously in their daily lives, instead of being collected in a controlled environment. Therefore, it offers new opportunities to understand the interrelationship among the personal, situational, and musical factors of music listening. As an example application, we present research investigating the interaction between the affective context of the listener and the affective content of music, using audio-based music emotion recognition techniques and a psycholinguistic tool. The study offers insights into the role of music in mood regulation and demonstrates how LJ2M can contribute to studies on real-world music listening behavior.

[1]  Jeffrey J. Scott,et al.  MUSIC EMOTION RECOGNITION: A STATE OF THE ART REVIEW , 2010 .

[2]  Bonnie A. Nardi,et al.  Why we blog , 2004, CACM.

[3]  C. Pelletier The effect of music on decreasing arousal due to stress: a meta-analysis. , 2004, Journal of music therapy.

[4]  Òscar Celma,et al.  Music Recommendation and Discovery - The Long Tail, Long Fail, and Long Play in the Digital Music Space , 2010 .

[5]  Philip R. Cohen,et al.  Tangible multimodal interfaces for safety-critical applications , 2004, CACM.

[6]  Joseph Kaye,et al.  Understanding how bloggers feel: recognizing affect in blog posts , 2006, CHI Extended Abstracts.

[7]  Thierry Bertin-Mahieux,et al.  The Million Song Dataset , 2011, ISMIR.

[8]  Boicho Kokinov,et al.  A Dynamic Theory of Implicit Context , 1997 .

[9]  Homer H. Chen,et al.  Music Emotion Recognition , 2011 .

[10]  M. Bradley,et al.  Affective Norms for English Words (ANEW): Instruction Manual and Affective Ratings , 1999 .

[11]  Francesco Ricci,et al.  Contextual music information retrieval and recommendation: State of the art and challenges , 2012, Comput. Sci. Rev..

[12]  Vaibhavi N Patodkar,et al.  Twitter as a Corpus for Sentiment Analysis and Opinion Mining , 2016 .

[13]  Preslav Nakov,et al.  SemEval-2013 Task 2: Sentiment Analysis in Twitter , 2013, *SEMEVAL.

[14]  K. Scherer Which Emotions Can be Induced by Music? What Are the Underlying Mechanisms? And How Can We Measure Them? , 2004 .

[15]  Òscar Celma Herrada Music recommendation and discovery in the long tail , 2009 .

[16]  Rafael A. Calvo,et al.  Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications , 2010, IEEE Transactions on Affective Computing.

[17]  Adrian C. North,et al.  Why do we listen to music? A uses and gratifications analysis. , 2011, British journal of psychology.

[18]  J. Pennebaker,et al.  Word Use in the Poetry of Suicidal and Nonsuicidal Poets , 2001, Psychosomatic medicine.

[19]  J. Pennebaker,et al.  Language use of depressed and depression-vulnerable college students , 2004 .

[20]  Yi-Hsuan Yang,et al.  Quantitative Study of Music Listening Behavior in a Social and Affective Context , 2013, IEEE Transactions on Multimedia.

[21]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[22]  J. Pennebaker,et al.  The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods , 2010 .

[23]  J. Sloboda,et al.  The functions of music for affect regulation , 2011 .

[24]  David S. Rosenblum,et al.  Context-aware mobile music recommendation for daily activities , 2012, ACM Multimedia.

[25]  Markus Schedl,et al.  The Million Musical Tweet Dataset - What We Can Learn From Microblogs , 2013, ISMIR.

[26]  Douglas Eck,et al.  The need for music information retrieval with user-centered and multimodal strategies , 2011, MIRUM '11.

[27]  Markus Schedl,et al.  Putting the User in the Center of Music Information Retrieval , 2012, ISMIR.