Paragraph-Level Commonsense Transformers with Recurrent Memory

Human understanding of narrative texts requires making commonsense inferences beyond what is stated in the text explicitly. A recent model, COMET, can generate such inferences along several dimensions such as pre- and post-conditions, motivations, and mental states of the participants. However, COMET was trained on short phrases, and is therefore discourse-agnostic. When presented with each sentence of a multi-sentence narrative, it might generate inferences that are inconsistent with the rest of the narrative. We present the task of discourse-aware commonsense inference. Given a sentence within a narrative, the goal is to generate commonsense inferences along predefined dimensions, while maintaining coherence with the rest of the narrative. Such large-scale paragraph-level annotation is hard to get and costly, so we use available sentence-level annotations to efficiently and automatically construct a distantly supervised corpus. Using this corpus, we train PARA-COMET, a discourse-aware model that incorporates paragraph-level information to generate coherent commonsense inferences from narratives. PARA-COMET captures both semantic knowledge pertaining to prior world knowledge, and episodic knowledge involving how current events relate to prior and future events in a narrative. Our results confirm that PARA-COMET outperforms the sentence-level baselines, particularly in generating inferences that are both coherent and novel.

[1]  Mark O. Riedl,et al.  Automated Storytelling via Causal, Commonsense Plot Ordering , 2020, AAAI.

[2]  Dong Si,et al.  A Wizard-of-Oz Interface and Persona-based Methodology for Collecting Health Counseling Dialog , 2020, CHI Extended Abstracts.

[3]  Minlie Huang,et al.  A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation , 2020, TACL.

[4]  Noah A. Smith,et al.  Social Bias Frames: Reasoning about Social and Power Implications of Language , 2019, ACL.

[5]  Lysandre Debut,et al.  HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.

[6]  R'emi Louf,et al.  HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.

[7]  Yejin Choi,et al.  Counterfactual Story Reasoning and Generation , 2019, EMNLP.

[8]  Dan Roth,et al.  “Going on a vacation” takes longer than “Going for a walk”: A Study of Temporal Commonsense Understanding , 2019, EMNLP.

[9]  Zhao Hai,et al.  Semantics-aware BERT for Language Understanding , 2019, AAAI.

[10]  Alexander M. Rush,et al.  Commonsense Knowledge Mining from Pretrained Models , 2019, EMNLP.

[11]  Yejin Choi,et al.  Cosmos QA: Machine Reading Comprehension with Contextual Commonsense Reasoning , 2019, EMNLP.

[12]  Doug Downey,et al.  Abductive Commonsense Reasoning , 2019, ICLR.

[13]  Yejin Choi,et al.  COMET: Commonsense Transformers for Automatic Knowledge Graph Construction , 2019, ACL.

[14]  Ting Liu,et al.  Constructing Narrative Event Evolutionary Graph for Script Event Prediction , 2018, IJCAI.

[15]  Yejin Choi,et al.  Modeling Naive Psychology of Characters in Simple Commonsense Stories , 2018, ACL.

[16]  Jackie Chi Kit Cheung,et al.  Commonsense mining as knowledge base completion? A study on the impact of novelty , 2018, ArXiv.

[17]  Catherine Havasi,et al.  ConceptNet 5.5: An Open Multilingual Graph of General Knowledge , 2016, AAAI.

[18]  Sheng Zhang,et al.  Ordinal Common-sense Inference , 2016, TACL.

[19]  Kyle A. Pettijohn,et al.  Narrative event boundaries, reading times, and expectation , 2016, Memory & cognition.

[20]  Nathanael Chambers,et al.  A Corpus and Cloze Evaluation for Deeper Understanding of Commonsense Stories , 2016, NAACL.

[21]  Stephen Clark,et al.  What Happens Next? Event Prediction Using a Compositional Neural Network Model , 2016, AAAI.

[22]  Francis Ferraro,et al.  Script Induction as Language Modeling , 2015, EMNLP.

[23]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[24]  Fabio Massimo Zanzotto,et al.  Book Reviews: Recognizing Textual Entailment: Models and Applications by Ido Dagan, Dan Roth, Mark Sammons and Fabio Massimo Zanzotto , 2013, CL.

[25]  Marie-Francine Moens,et al.  Skip N-grams and Ranking Functions for Predicting Script Events , 2012, EACL.

[26]  Zornitsa Kozareva,et al.  Learning Temporal Information for States and Events , 2011, 2011 IEEE Fifth International Conference on Semantic Computing.

[27]  Zornitsa Kozareva,et al.  SemEval-2012 Task 7: Choice of Plausible Alternatives: An Evaluation of Commonsense Causal Reasoning , 2011, *SEMEVAL.

[28]  Nathanael Chambers,et al.  Unsupervised Learning of Narrative Event Chains , 2008, ACL.

[29]  Chin-Yew Lin,et al.  ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.

[30]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[31]  John Hale,et al.  A Probabilistic Earley Parser as a Psycholinguistic Model , 2001, NAACL.

[32]  Roger C. Schank,et al.  Scripts, plans, goals and understanding: an inquiry into human knowledge structures , 1978 .

[33]  Terry Winograd,et al.  Understanding natural language , 1974 .

[34]  Eugene Charniak,et al.  Toward a model of children's story comprehension , 1972 .

[35]  Debanjan Ghosh,et al.  R3: Reverse, Retrieve, and Rank for Sarcasm Generation with Commonsense Knowledge , 2020, ACL.

[36]  Yejin Choi,et al.  ATOMIC: An Atlas of Machine Commonsense for If-Then Reasoning , 2019, AAAI.

[37]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[38]  Ilya Sutskever,et al.  Language Models are Unsupervised Multitask Learners , 2019 .

[39]  Rémi Louf,et al.  Transformers : State-ofthe-art Natural Language Processing , 2019 .

[40]  Alec Radford,et al.  Improving Language Understanding by Generative Pre-Training , 2018 .

[41]  Henry Lieberman,et al.  Understanding Stories with Large-Scale Common Sense , 2017, COMMONSENSE.

[42]  E. Tulving,et al.  Episodic and semantic memory , 1972 .

[43]  Raymond J. Mooney,et al.  Statistical Script Learning with Multi-Argument Events , 2014, EACL.

[44]  Sang Joon Kim,et al.  A Mathematical Theory of Communication , 2006 .

[45]  J. Fleiss Measuring nominal scale agreement among many raters. , 1971 .

[46]  F. Heider The psychology of interpersonal relations , 1958 .