PEIA: Personality and Emotion Integrated Attentive Model for Music Recommendation on Social Media Platforms

With the rapid expansion of digital music formats, it's indispensable to recommend users with their favorite music. For music recommendation, users' personality and emotion greatly affect their music preference, respectively in a long-term and short-term manner, while rich social media data provides effective feedback on these information. In this paper, aiming at music recommendation on social media platforms, we propose a Personality and Emotion Integrated Attentive model (PEIA), which fully utilizes social media data to comprehensively model users' long-term taste (personality) and short-term preference (emotion). Specifically, it takes full advantage of personality-oriented user features, emotion-oriented user features and music features of multi-faceted attributes. Hierarchical attention is employed to distinguish the important factors when incorporating the latent representations of users' personality and emotion. Extensive experiments on a large real-world dataset of 171,254 users demonstrate the effectiveness of our PEIA model which achieves an NDCG of 0.5369, outperforming the state-of-the-art methods. We also perform detailed parameter analysis and feature contribution analysis, which further verify our scheme and demonstrate the significance of co-modeling of user personality and emotion in music recommendation.

[1]  D. Funder,et al.  Emotional experience in daily life: valence, variability, and rate of change. , 2001, Emotion.

[2]  John Riedl,et al.  Item-based collaborative filtering recommendation algorithms , 2001, WWW '01.

[3]  S. Gosling,et al.  PERSONALITY PROCESSES AND INDIVIDUAL DIFFERENCES The Do Re Mi’s of Everyday Life: The Structure and Personality Correlates of Music Preferences , 2003 .

[4]  Ichiro Fujinaga,et al.  Combining Features Extracted from Audio, Symbolic and Cultural Sources , 2008, ISMIR.

[5]  Steffen Rendle,et al.  Factorization Machines , 2010, 2010 IEEE International Conference on Data Mining.

[6]  Thierry Bertin-Mahieux,et al.  The Million Song Dataset , 2011, ISMIR.

[7]  Adrian C. North,et al.  Why do we listen to music? A uses and gratifications analysis. , 2011, British journal of psychology.

[8]  Y. Song,et al.  A Survey of Music Recommendation Systems and Future Perspectives , 2012 .

[9]  Björn W. Schuller,et al.  Recent developments in openSMILE, the munich open-source multimedia feature extractor , 2013, ACM Multimedia.

[10]  Benjamin Schrauwen,et al.  Deep content-based music recommendation , 2013, NIPS.

[11]  Quoc V. Le,et al.  Distributed Representations of Sentences and Documents , 2014, ICML.

[12]  Guandong Xu,et al.  Exploring user emotion in microblogs for music recommendation , 2015, Expert Syst. Appl..

[13]  Tao Chen,et al.  TriRank: Review-aware Explainable Recommendation by Modeling Aspects , 2015, CIKM.

[14]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[15]  Björn W. Schuller,et al.  Emotion in the singing voice—a deeperlook at acoustic features in the light ofautomatic classification , 2015, EURASIP J. Audio Speech Music. Process..

[16]  Bruce Ferwerda,et al.  Personality & Emotional States: Understanding Users' Music Listening Needs , 2015, UMAP Workshops.

[17]  Matthijs Douze,et al.  FastText.zip: Compressing text classification models , 2016, ArXiv.

[18]  Jun Wang,et al.  Product-Based Neural Networks for User Response Prediction , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).

[19]  Heng-Tze Cheng,et al.  Wide & Deep Learning for Recommender Systems , 2016, DLRS@RecSys.

[20]  Rui Cheng,et al.  A Music Recommendation System Based on Acoustic Features and User Personalities , 2016, PAKDD Workshops.

[21]  Gang Fu,et al.  Deep & Cross Network for Ad Click Predictions , 2017, ADKDD@KDD.

[22]  Yunming Ye,et al.  DeepFM: A Factorization-Machine based Neural Network for CTR Prediction , 2017, IJCAI.

[23]  Mohan S. Kankanhalli,et al.  Exploring User-Specific Information in Music Retrieval , 2017, SIGIR.

[24]  Xin Zhang,et al.  Learning to embed music and metadata for context-aware music recommendation , 2018, World Wide Web.

[25]  Hanwang Zhang,et al.  Filtering : Multimedia Recommendation with Item-and Component-Level A ention , 2017 .

[26]  Tat-Seng Chua,et al.  Attentional Factorization Machines: Learning the Weight of Feature Interactions via Attention Networks , 2017, IJCAI.

[27]  Vikram Pudi,et al.  Attentive neural architecture incorporating song features for music recommendation , 2018, RecSys.

[28]  Keiichiro Hoashi,et al.  Mood-Aware Music Recommendation via Adaptive Song Embedding , 2018, 2018 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM).

[29]  Xing Xie,et al.  xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems , 2018, KDD.

[30]  Xiangnan He,et al.  NAIS: Neural Attentive Item Similarity Model for Recommendation , 2018, IEEE Transactions on Knowledge and Data Engineering.

[31]  Guorui Zhou,et al.  Deep Interest Network for Click-Through Rate Prediction , 2017, KDD.

[32]  Juhan Nam,et al.  Deep Content-User Embedding Model for Music Recommendation , 2018, ArXiv.