The optimally designed dynamic memory networks for targeted sentiment classification

Abstract This paper focuses on the task of targeted sentiment classification over dynamic memory networks. To remedy the flaw of the previous related work, only doing well in simple sentences, the new dynamic memory networks, in which the input module, question module and memory module are optimally designed, are established to model the task into question answering system for the first time. Specifically, in the input module, position embedding is considered to eliminate the influence of sentiment-irrelevant words that close to the target, and skip connection can help to build rich representations over the sentences. In the question module, for dealing with many cases where the target is composed of multiple words, the commonly used average target vector is replaced by the designed target sentiment question, which is encoded by a GRU. In the memory module, two improvements are performed. On the one hand, weight bias, caused by using original soft attention to extract memory information at each attention step, is avoided by both exploring attention and inner attention based GRU networks. On the other hand, memory updating block is capable to distill sentiment-relevant features from memory information, which is extracted by multiple attention block using multiple attention mechanism. Finally, experimental results demonstrate the effectiveness of the proposed models for the task of targeted sentiment classification among three different datasets. Additionally, the best model of this paper achieves state-of-the-art performance on different types of data.

[1]  Yue Zhang,et al.  Target-Dependent Twitter Sentiment Classification with Rich Automatic Features , 2015, IJCAI.

[2]  Yann Dauphin,et al.  Convolutional Sequence to Sequence Learning , 2017, ICML.

[3]  Saif Mohammad,et al.  NRC-Canada-2014: Detecting Aspects and Sentiment in Customer Reviews , 2014, *SEMEVAL.

[4]  Dong-Hong Ji,et al.  Towards Twitter sentiment classification by multi-level sentiment-enriched word embeddings , 2016, Neurocomputing.

[5]  Stefan Wermter,et al.  Emotion-modulated attention improves expression recognition: A deep learning model , 2017, Neurocomputing.

[6]  Arkaitz Zubiaga,et al.  TDParse: Multi-target-specific sentiment recognition on Twitter , 2017, EACL.

[7]  Yue Zhang,et al.  Context-Sensitive Lexicon Features for Neural Sentiment Analysis , 2016, EMNLP.

[8]  Lidong Bing,et al.  Recurrent Attention Network on Memory for Aspect Sentiment Analysis , 2017, EMNLP.

[9]  Yidong Chen,et al.  Sentiment analysis via integrating distributed representations of variable-length word sequence , 2016, Neurocomputing.

[10]  Irwin King,et al.  Aspect-level Sentiment Classification with HEAT (HiErarchical ATtention) Network , 2017, CIKM.

[11]  Yue Zhang,et al.  Neural Networks for Open Domain Targeted Sentiment , 2015, EMNLP.

[12]  Yue Zhang,et al.  Attention Modeling for Targeted Sentiment , 2017, EACL.

[13]  Li Zhao,et al.  Attention-based LSTM for Aspect-level Sentiment Classification , 2016, EMNLP.

[14]  Cheng Li,et al.  Deep Memory Networks for Attitude Identification , 2017, WSDM.

[15]  Richard Socher,et al.  Dynamic Coattention Networks For Question Answering , 2016, ICLR.

[16]  Siu Cheung Hui,et al.  Dyadic Memory Networks for Aspect-based Sentiment Analysis , 2017, CIKM.

[17]  Tiejun Zhao,et al.  Target-dependent Twitter Sentiment Classification , 2011, ACL.

[18]  Suresh Manandhar,et al.  SemEval-2015 Task 12: Aspect Based Sentiment Analysis , 2015, *SEMEVAL.

[19]  Richard Socher,et al.  Dynamic Memory Networks for Visual and Textual Question Answering , 2016, ICML.

[20]  Kim Schouten,et al.  Survey on Aspect-Level Sentiment Analysis , 2016, IEEE Transactions on Knowledge and Data Engineering.

[21]  Zhiyuan Liu,et al.  Relation Classification via Multi-Level Attention CNNs , 2016, ACL.

[22]  Ming Zhou,et al.  Adaptive Recursive Neural Network for Target-dependent Twitter Sentiment Classification , 2014, ACL.

[23]  Richard Socher,et al.  Ask Me Anything: Dynamic Memory Networks for Natural Language Processing , 2015, ICML.

[24]  Min Yang,et al.  Attention Based LSTM for Target Dependent Sentiment Classification , 2017, AAAI.

[25]  Ting Liu,et al.  Aspect Level Sentiment Classification with Deep Memory Network , 2016, EMNLP.

[26]  Xiaocheng Feng,et al.  Effective LSTMs for Target-Dependent Sentiment Classification , 2015, COLING.

[27]  Laizhong Cui,et al.  Combine HowNet lexicon to train phrase recursive autoencoder for sentence-level sentiment analysis , 2017, Neurocomputing.

[28]  Saif Mohammad,et al.  Sentiment Analysis of Short Informal Texts , 2014, J. Artif. Intell. Res..

[29]  Houfeng Wang,et al.  Interactive Attention Networks for Aspect-Level Sentiment Classification , 2017, IJCAI.

[30]  Suresh Manandhar,et al.  SemEval-2014 Task 4: Aspect Based Sentiment Analysis , 2014, *SEMEVAL.

[31]  Alexander J. Smola,et al.  Stacked Attention Networks for Image Question Answering , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  Haris Papageorgiou,et al.  SemEval-2016 Task 5: Aspect Based Sentiment Analysis , 2016, *SEMEVAL.

[33]  Yang Zou,et al.  Textual sentiment analysis via three different attention convolutional neural networks and cross-modality consistent regression , 2018, Neurocomputing.

[34]  Yue Zhang,et al.  Gated Neural Networks for Targeted Sentiment Analysis , 2016, AAAI.

[35]  Jun Zhao,et al.  Inner Attention based Recurrent Neural Networks for Answer Selection , 2016, ACL.