A Multi-Task Learning Approach to Improve Sentiment Analysis with Explicit Recommendation

When expressing sentiment towards products, customers often explicitly indicate their recommendation status. Nevertheless, most existing literature focuses on sentiment analysis but neglects the rich correlation information that may be brought by explicit recommendation classification. We argue that the two tasks are correlated and hence, the knowledge in explicit recommendation classification can also be beneficial to sentiment analysis. Consequently, in this paper, a novel bidirectional encoder representations from transformers (BERT)-enhanced multi-task learning (BeMTL) approach is proposed to improve sentiment analysis with explicit recommendation classification. Specifically, the proposed MTL approach takes contextualized word embeddings produced by the pre-trained BERT-based embedding layer. Then, it learns the sentence contextual features shared between both tasks with a convolutional multi-head attention neural network. To fully exploit the correlation information between sentiment analysis and explicit recommendation classification tasks, a novel inter-task matching layer (IML) is designed to match their representations. In nutshell, our study reveals the potential of multi-task learning models on such types of problems, and experimental results on two Amazon datasets show that our approach outperforms the state-of-the-art baseline approaches for sentiment analysis.

[1]  Yaohui Jin,et al.  Multi-Task Label Embedding for Text Classification , 2017, EMNLP.

[2]  Yann LeCun,et al.  Very Deep Convolutional Networks for Text Classification , 2016, EACL.

[3]  Tong Zhang,et al.  Deep Pyramid Convolutional Neural Networks for Text Categorization , 2017, ACL.

[4]  Tao Shen,et al.  DiSAN: Directional Self-Attention Network for RNN/CNN-free Language Understanding , 2017, AAAI.

[5]  Massimiliano Pontil,et al.  The Benefit of Multitask Representation Learning , 2015, J. Mach. Learn. Res..

[6]  Lukasz Kaiser,et al.  One Model To Learn Them All , 2017, ArXiv.

[7]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[8]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[9]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[10]  Richard Socher,et al.  Dynamic Coattention Networks For Question Answering , 2016, ICLR.

[11]  Rich Caruana,et al.  Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.

[12]  Björn W. Schuller,et al.  Contextual Bidirectional Long Short-Term Memory Recurrent Neural Network Language Models: A Generative Approach to Sentiment Analysis , 2017, EACL.

[13]  Diyi Yang,et al.  Hierarchical Attention Networks for Document Classification , 2016, NAACL.

[14]  Kang Liu,et al.  Book Review: Sentiment Analysis: Mining Opinions, Sentiments, and Emotions by Bing Liu , 2015, CL.

[15]  Abien Fred Agarap,et al.  Statistical Analysis on E-Commerce Reviews, with Sentiment Classification using Bidirectional Recurrent Neural Network (RNN) , 2018, ArXiv.

[16]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[17]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[18]  Huajie Chen,et al.  Multi-Task Multi-Head Attention Memory Network for Fine-Grained Sentiment Analysis , 2019, NLPCC.

[19]  Yue Yin,et al.  Explainable Recommendation via Multi-Task Learning in Opinionated Text Data , 2018, SIGIR.

[20]  Zhen Wu,et al.  Improving Review Representations with User Attention and Product Attention for Sentiment Classification , 2018, AAAI.

[21]  Jonah Berger,et al.  How Language Shapes Word of Mouth's Impact , 2017 .

[22]  Quoc V. Le,et al.  QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension , 2018, ICLR.

[23]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[24]  Barry Smyth,et al.  Why I like it: multi-task learning for recommendation and explanation , 2018, RecSys.

[25]  Georgios Balikas,et al.  Multitask Learning for Fine-Grained Twitter Sentiment Analysis , 2017, SIGIR.

[26]  Ge Yu,et al.  Sentiment analysis using deep learning approaches: an overview , 2019, Science China Information Sciences.

[27]  Paulo Rita,et al.  How to predict explicit recommendations in online reviews using text mining and sentiment analysis , 2020, Journal of Hospitality and Tourism Management.

[28]  Ye Zhang,et al.  A Sensitivity Analysis of (and Practitioners’ Guide to) Convolutional Neural Networks for Sentence Classification , 2015, IJCNLP.

[29]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[30]  Erik Cambria,et al.  Sentiment and Sarcasm Classification With Multitask Learning , 2019, IEEE Intelligent Systems.

[31]  Zhaopeng Tu,et al.  Convolutional Self-Attention Networks , 2019, NAACL.