Exploring Fine-Tuned Embeddings that Model Intensifiers for Emotion Analysis

Adjective phrases like "a little bit surprised", "completely shocked", or "not stunned at all" are not handled properly by currently published state-of-the-art emotion classification and intensity prediction systems which use pre-dominantly non-contextualized word embeddings as input. Based on this finding, we analyze differences between embeddings used by these systems in regard to their capability of handling such cases. Furthermore, we argue that intensifiers in context of emotion words need special treatment, as is established for sentiment polarity classification, but not for more fine-grained emotion prediction. To resolve this issue, we analyze different aspects of a post-processing pipeline which enriches the word representations of such phrases. This includes expansion of semantic spaces at the phrase level and sub-word level followed by retrofitting to emotion lexica. We evaluate the impact of these steps with A La Carte and Bag-of-Substrings extensions based on pretrained GloVe, Word2vec, and fastText embeddings against a crowd-sourced corpus of intensity annotations for tweets containing our focus phrases. We show that the fastText-based models do not gain from handling these specific phrases under inspection. For Word2vec embeddings, we show that our post-processing pipeline improves the results by up to 8% on a novel dataset densely populated with intensifiers.

[1]  Sali A. Tagliamonte ''So cool, right?'': Canadian English Entering the 21st Century , 2006 .

[2]  Saif Mohammad,et al.  Sentiment Composition of Words with Opposing Polarities , 2016, NAACL.

[3]  Jordan J. Louviere,et al.  Best-Worst Scaling: Theory, Methods and Applications , 2015 .

[4]  Jan Svartvik,et al.  A __ comprehensive grammar of the English language , 1988 .

[5]  A. Athanasiadou On the subjectivity of intensifiers , 2007 .

[6]  T. Nevalainen Social variation in intensifier use: constraint on -ly adverbialization in the past?1 , 2008, English Language and Linguistics.

[7]  Iyad Rahwan,et al.  Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm , 2017, EMNLP.

[8]  Eugene H. Ehrlich Oxford American Dictionary , 1980 .

[9]  Isaac G. Councill,et al.  What's great and what's not: learning to classify the scope of negation for improved sentiment analysis , 2010, NeSp-NLP@ACL.

[10]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[11]  Carita Paradis Adjectives and boundedness , 2001 .

[12]  Maria Napoli,et al.  Exploring Intensification: Synchronic, diachronic and cross-linguistic perspectives , 2017 .

[13]  Mikhail Khodak,et al.  A La Carte Embedding: Cheap but Effective Induction of Semantic Feature Vectors , 2018, ACL.

[14]  Raffaella Bernardi,et al.  A Distributional Study of Negated Adjectives and Antonyms , 2018, CLiC-it.

[15]  Roman Klinger,et al.  An Empirical Analysis of the Role of Amplifiers, Downtoners, and Negations in Emotion Classification in Microblogs , 2018, 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA).

[16]  Roser Morante,et al.  Modality and Negation: An Introduction to the Special Issue , 2012, CL.

[17]  Jorge Carrillo de Albornoz,et al.  An emotion-based model of negation, intensifiers, and modality for polarity and intensity classification , 2013, J. Assoc. Inf. Sci. Technol..

[18]  Yingyu Liang,et al.  Generalizing Word Embeddings using Bag of Subwords , 2018, EMNLP.

[19]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[20]  Hongyu Guo,et al.  An Empirical Study on the Effect of Negation Words on Sentiment , 2014, ACL.

[21]  Saif Mohammad,et al.  Capturing Reliable Fine-Grained Sentiment Associations by Crowdsourcing and Best–Worst Scaling , 2016, NAACL.

[22]  Maria Napoli,et al.  New insights on intensification and intensifiers , 2017 .

[23]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[24]  Paloma Núñez Pertejo,et al.  That’s absolutely crap, totally rubbish : The use of the intensifiers absolutely and totally in the spoken language of British adults and teenagers , 2014 .

[25]  P. Ekman An argument for basic emotions , 1992 .

[26]  T. Nevalainen,et al.  Fairly pretty or pretty fair? On the development and grammaticalization of English downtoners , 2002 .

[27]  Saif Mohammad,et al.  SemEval-2018 Task 1: Affect in Tweets , 2018, *SEMEVAL.

[28]  Degree adverbs in Early Modern English , 1994 .

[29]  Carita Paradis,et al.  It’s well weird: Degree Modifiers of Adjectives Revisited: The Nineties , 2000, Corpora Galore.

[30]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[31]  Christiane Fellbaum,et al.  The Role of Adverbs in Sentiment Analysis , 2014 .

[32]  Luke S. Zettlemoyer,et al.  Deep Contextualized Word Representations , 2018, NAACL.

[33]  Björn Gambäck,et al.  Negation Scope Detection for Twitter Sentiment Analysis , 2015, WASSA@EMNLP.

[34]  Tomas Mikolov,et al.  Enriching Word Vectors with Subword Information , 2016, TACL.

[35]  Carita Paradis Degree Modifiers of Adjectives in Spoken British English , 1997 .

[36]  Saif M. Mohammad,et al.  The Effect of Negators, Modals, and Degree Adverbs on Sentiment Composition , 2016, WASSA@NAACL-HLT.

[37]  David Vandyke,et al.  Counter-fitting Word Vectors to Linguistic Constraints , 2016, NAACL.

[38]  Saif Mohammad,et al.  WASSA-2017 Shared Task on Emotion Intensity , 2017, WASSA@EMNLP.

[39]  R. Plutchik A GENERAL PSYCHOEVOLUTIONARY THEORY OF EMOTION , 1980 .

[40]  J. Russell,et al.  The circumplex model of affect: An integrative approach to affective neuroscience, cognitive development, and psychopathology , 2005, Development and Psychopathology.

[41]  Chuhan Wu,et al.  THU_NGN at SemEval-2018 Task 1: Fine-grained Tweet Sentiment Intensity Analysis with Attention CNN-LSTM , 2018, *SEMEVAL.