SNNN: Promoting Word Sentiment and Negation in Neural Sentiment Classification

We mainly investigate word influence in neural sentiment classification, which results in a novel approach to promoting word sentiment and negation as attentions. Particularly, a sentiment and negation neural network (SNNN) is proposed, including a sentiment neural network (SNN) and a negation neural network (NNN). First, we modify the word level by embedding the word sentiment and negation information as the extra layers for the input. Second, we adopt a hierarchical LSTM model to generate the word-level, sentence-level and document-level representations respectively. After that, we enhance word sentiment and negation as attentions over the semantic level. Finally, the experiments conducting on the IMDB and Yelp data sets show that our approach is superior to the state-of-the-art methods. Furthermore, we draw the interesting conclusions that (1) LSTM performs better than CNN and RNN for neural sentiment classification; (2) word sentiment and negation are a strong alliance as attentions, while overfitting occurs when they are simultaneously applied at the embedding layer; and (3) word sentiment/negation can be singly implemented for better performance as both embedding layer and attention at the same time. Introduction and Motivation Many approaches in sentiment classification, utilize a supervised classifier and rely on extensive feature engineering (Go, Bhayani, and Huang 2009; Barbosa and Feng 2010; Pak and Paroubek 2010; Jiang et al. 2011; Mukherjee, Bhattacharyya, and others 2012; Hamdan, Béchet, and Bellot 2013; Mohammad, Kiritchenko, and Zhu 2013; Cheng et al. 2017). However, feature engineering costs extensive labour work and needs specific domain knowledge. Therefore, feature learning is an alternative way to learn discriminative features automatically from data. The work presented by (Socher et al. 2013; Yessenalina and Cardie 2011; Hu et al. 2016) proved that the features of a sentence/document could be learnt through its word embedding. Existing approaches of learning word embedding (Collobert et al. 2011; Mikolov et al. 2013b; Yang, Hu, and He 2015) then focused on modeling the syntactic context. After that, people turn to neural network for its learning ability of text representation (Glorot, Bordes, and Bengio 2011; Zhai and Zhang Copyright c © 2018, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 2016; Socher et al. 2011a; 2011b; 2012; 2013; Kim 2014; Tang, Qin, and Liu 2015b; Yang et al. 2016; Chen et al. 2016; Ren et al. 2016a) Tang et al. (Tang, Qin, and Liu 2015b) proposed a neural network model to learn vector-based document representation in a CNN based sentiment classification, where the authors found that neural gates outperformed the traditional recurrent neural network. Then, Chen et al. (Chen et al. 2016) brought a hierarchical neural network to incorporate global user and product information as attentions. They mainly challenged Tang’s work that the characteristics of the user and product information should be reflected on the semantic level, instead of the word level. Based on these two pieces of the state-ofthe-art work, we aim to investigate further word influence in neural sentiment classification. The motivation is that it is theoretically feasible and sound by adding more word information from multiple dimensions in the word level as the input, since the quality of document/sentence representation highly depends on word representations. I positive it not negative output it

[1]  Yoshua Bengio,et al.  Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.

[2]  Ming Zhou,et al.  Learning Sentiment-Specific Word Embedding for Twitter Sentiment Classification , 2014, ACL.

[3]  Alexander J. Smola,et al.  Jointly modeling aspects, ratings and sentiments for movie recommendation (JMARS) , 2014, KDD.

[4]  Qinmin Hu,et al.  SG++: Word Representation with Sentiment and Negation for Twitter Sentiment Classification , 2016, SIGIR.

[5]  Qinmin Hu,et al.  Enhancing Recurrent Neural Networks with Positional Attention for Question Answering , 2017, SIGIR.

[6]  Vaibhavi N Patodkar,et al.  Twitter as a Corpus for Sentiment Analysis and Opinion Mining , 2016 .

[7]  Annie Zaenen,et al.  Contextual Valence Shifters , 2006, Computing Attitude and Affect in Text.

[8]  Quoc V. Le,et al.  Distributed Representations of Sentences and Documents , 2014, ICML.

[9]  Christopher Potts,et al.  Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.

[10]  Jeffrey Pennington,et al.  Semi-Supervised Recursive Autoencoders for Predicting Sentiment Distributions , 2011, EMNLP.

[11]  Bo Pang,et al.  Thumbs up? Sentiment Classification using Machine Learning Techniques , 2002, EMNLP.

[12]  Qinmin Hu,et al.  ECNU at TREC 2015: Microblog Track , 2015, TREC.

[13]  Saif Mohammad,et al.  NRC-Canada: Building the State-of-the-Art in Sentiment Analysis of Tweets , 2013, *SEMEVAL.

[14]  Min Yang,et al.  Attention Based LSTM for Target Dependent Sentiment Classification , 2017, AAAI.

[15]  Junlan Feng,et al.  Robust Sentiment Detection on Twitter from Biased and Noisy Data , 2010, COLING.

[16]  Ting Liu,et al.  Learning Semantic Representations of Users and Products for Document Level Sentiment Classification , 2015, ACL.

[17]  Stephanie Seneff,et al.  Review Sentiment Scoring via a Parse-and-Paraphrase Paradigm , 2009, EMNLP.

[18]  Claire Cardie,et al.  Compositional Matrix-Space Models for Sentiment Analysis , 2011, EMNLP.

[19]  Zhongfei Zhang,et al.  Structural Correspondence Learning for Cross-Lingual Sentiment Classification with One-to-Many Mappings , 2016, AAAI.

[20]  Andrew Y. Ng,et al.  Semantic Compositionality through Recursive Matrix-Vector Spaces , 2012, EMNLP.

[21]  Hongliang Yu,et al.  Identifying Sentiment Words Using an Optimization Model with L1 Regularization , 2016, AAAI.

[22]  Diyi Yang,et al.  Hierarchical Attention Networks for Document Classification , 2016, NAACL.

[23]  Yoshua Bengio,et al.  A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..

[24]  Qinmin Hu,et al.  Locating Query-oriented Experts in Microblog Search , 2014, SMIR@SIGIR.

[25]  Zhiyuan Liu,et al.  Neural Sentiment Classification with User and Product Attention , 2016, EMNLP.

[26]  Yue Zhang,et al.  Context-Sensitive Twitter Sentiment Classification Using Neural Network , 2016, AAAI.

[27]  Yue Zhang,et al.  Improving Twitter Sentiment Classification Using Topic-Enriched Multi-Prototype Word Embeddings , 2016, AAAI.

[28]  Saif Mohammad,et al.  Sentiment Analysis of Short Informal Texts , 2014, J. Artif. Intell. Res..

[29]  Alistair Kennedy,et al.  SENTIMENT CLASSIFICATION of MOVIE REVIEWS USING CONTEXTUAL VALENCE SHIFTERS , 2006, Comput. Intell..

[30]  Yoon Kim,et al.  Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.

[31]  Tiejun Zhao,et al.  Target-dependent Twitter Sentiment Classification , 2011, ACL.

[32]  Jason Weston,et al.  Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..

[33]  Andrew Y. Ng,et al.  Parsing Natural Scenes and Natural Language with Recursive Neural Networks , 2011, ICML.

[34]  Ting Liu,et al.  Document Modeling with Gated Recurrent Neural Network for Sentiment Classification , 2015, EMNLP.

[35]  Maite Taboada,et al.  Lexicon-Based Methods for Sentiment Analysis , 2011, CL.

[36]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[37]  Tao Chen,et al.  Expert Systems With Applications , 2022 .

[38]  Qinmin Hu,et al.  Learning Topic-Oriented Word Embedding for Query Classification , 2015, PAKDD.

[39]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[40]  Zhongfei Zhang,et al.  Semisupervised Autoencoder for Sentiment Analysis , 2015, AAAI.

[41]  Hongyu Guo,et al.  An Empirical Study on the Effect of Negation Words on Sentiment , 2014, ACL.

[42]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[43]  Pushpak Bhattacharyya,et al.  Sentiment Analysis in Twitter with Lightweight Discourse Analysis , 2012, COLING.

[44]  Frédéric Béchet,et al.  Experiments with DBpedia, WordNet and SentiWordNet as resources for sentiment analysis in micro-blogging , 2013, *SEMEVAL.

[45]  Kewei Cheng,et al.  Unsupervised Sentiment Analysis with Signed Social Networks , 2017, AAAI.