Seeing Things from a Different Angle:Discovering Diverse Perspectives about Claims

One key consequence of the information revolution is a significant increase and a contamination of our information supply. The practice of fact checking won’t suffice to eliminate the biases in text data we observe, as the degree of factuality alone does not determine whether biases exist in the spectrum of opinions visible to us. To better understand controversial issues, one needs to view them from a diverse yet comprehensive set of perspectives. For example, there are many ways to respond to a claim such as “animals should have lawful rights”, and these responses form a spectrum of perspectives, each with a stance relative to this claim and, ideally, with evidence supporting it. Inherently, this is a natural language understanding task, and we propose to address it as such. Specifically, we propose the task of substantiated perspective discovery where, given a claim, a system is expected to discover a diverse set of well-corroborated perspectives that take a stance with respect to the claim. Each perspective should be substantiated by evidence paragraphs which summarize pertinent results and facts. We construct PERSPECTRUM, a dataset of claims, perspectives and evidence, making use of online debate websites to create the initial data collection, and augmenting it using search engines in order to expand and diversify our dataset. We use crowd-sourcing to filter out noise and ensure high-quality data. Our dataset contains 1k claims, accompanied with pools of 10k and 8k perspective sentences and evidence paragraphs, respectively. We provide a thorough analysis of the dataset to highlight key underlying language understanding challenges, and show that human baselines across multiple subtasks far outperform ma-chine baselines built upon state-of-the-art NLP techniques. This poses a challenge and opportunity for the NLP community to address.

[1]  Jacob Cohen,et al.  The Equivalence of Weighted Kappa and the Intraclass Correlation Coefficient as Measures of Reliability , 1973 .

[2]  J. Heckman Sample selection bias as a specification error , 1979 .

[3]  H. Markovits,et al.  The belief-bias effect in the production and evaluation of logical conclusions , 1989, Memory & cognition.

[4]  Simone Teufel,et al.  Argumentative zoning information extraction from scientific text , 1999 .

[5]  Chris Callison-Burch,et al.  Paraphrasing with Bilingual Parallel Corpora , 2005, ACL.

[6]  Marie-Francine Moens,et al.  Argumentation mining: the detection, classification and structure of arguments in text , 2009, ICAIL.

[7]  Ido Dagan,et al.  The Sixth PASCAL Recognizing Textual Entailment Challenge , 2009, TAC.

[8]  Dan Roth,et al.  Knowing What to Believe (when you already know something) , 2010, COLING.

[9]  Peter Clark,et al.  The Seventh PASCAL Recognizing Textual Entailment Challenge , 2011, TAC.

[10]  Graeme Hirst,et al.  Classifying arguments by scheme , 2011, ACL.

[11]  James B. Freeman,et al.  Argument Structure: Representation and Theory , 2011, Argumentation Library.

[12]  Serena Villata,et al.  Combining Textual Entailment and Argumentation Theory for Supporting Online Debates Interactions , 2012, ACL.

[13]  Floris Bex,et al.  Implementing the argument web , 2013, Commun. ACM.

[14]  Dan Roth,et al.  Latent credibility analysis , 2013, WWW.

[15]  Ido Dagan,et al.  Recognizing Textual Entailment: Models and Applications , 2013, Recognizing Textual Entailment: Models and Applications.

[16]  D. Roth,et al.  Judging the Veracity of Claims and Reliability of Sources with Fact-Finders , 2014 .

[17]  Noam Slonim,et al.  Context Dependent Claim Detection , 2014, COLING.

[18]  Jan Snajder,et al.  Back up your Stance: Recognizing Arguments in Online Discussions , 2014, ArgMining@ACL.

[19]  Vincent Ng,et al.  Automatic Keyphrase Extraction: A Survey of the State of the Art , 2014, ACL.

[20]  Vincent Ng,et al.  Why are You Taking this Stance? Identifying and Classifying Reasons in Ideological Debates , 2014, EMNLP.

[21]  Claire Cardie,et al.  Identifying Appropriate Support for Propositions in Online User Comments , 2014, ArgMining@ACL.

[22]  Floris Bex,et al.  ArguBlogging: An application for the Argument Web , 2014, J. Web Semant..

[23]  Noam Slonim,et al.  A Benchmark Dataset for Automatic Detection of Claims and Evidence in the Context of Controversial Topics , 2014, ArgMining@ACL.

[24]  Andreas Vlachos,et al.  Fact Checking: Task definition and dataset construction , 2014, LTCSS@ACL.

[25]  Mitesh M. Khapra,et al.  Show Me Your Evidence - an Automatic Method for Context Dependent Evidence Detection , 2015, EMNLP.

[26]  Eric Gilbert,et al.  CREDBANK: A Large-Scale Social Media Corpus With Associated Credibility Annotations , 2015, ICWSM.

[27]  Brian Ecker,et al.  Argument Mining: Extracting Arguments from Online Dialogue , 2015, SIGDIAL Conference.

[28]  Dan Roth,et al.  Overcoming bias to learn about controversial topics , 2015, J. Assoc. Inf. Sci. Technol..

[29]  Paolo Torroni,et al.  Argument Mining from Speech: Detecting Claims in Political Debates , 2016, AAAI.

[30]  Oren Etzioni,et al.  Combining Retrieval, Statistics, and Inference to Answer Elementary Science Questions , 2016, AAAI.

[31]  Manuela M. Veloso,et al.  ClaimEval: Integrated and Flexible Framework for Claim Evaluation Using Credibility of Sources , 2016, AAAI.

[32]  Karin Baier,et al.  The Uses Of Argument , 2016 .

[33]  Andreas Vlachos,et al.  Emergent: a novel data-set for stance classification , 2016, NAACL.

[34]  Matthias Hagen,et al.  A News Editorial Corpus for Mining Argumentation Strategies , 2016, COLING.

[35]  Saif Mohammad,et al.  SemEval-2016 Task 6: Detecting Stance in Tweets , 2016, *SEMEVAL.

[36]  William Yang Wang “Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection , 2017, ACL.

[37]  Diana Inkpen,et al.  A Dataset for Multi-Target Stance Detection , 2017, EACL.

[38]  Eugenio Tacchini,et al.  Some Like it Hoax: Automated Fake News Detection in Social Networks , 2017, ArXiv.

[39]  Fan Zhang,et al.  A Corpus of Annotated Revisions for Studying Argumentative Writing , 2017, ACL.

[40]  Akiko Aizawa,et al.  Prerequisite Skills for Reading Comprehension: Multi-Perspective Analysis of MCTest Datasets and Systems , 2017, AAAI.

[41]  Indrajit Bhattacharya,et al.  Stance Classification of Context-Dependent Claims , 2017, EACL.

[42]  Iryna Gurevych,et al.  Argumentation Mining in User-Generated Web Discourse , 2016, CL.

[43]  Xinyu Hua,et al.  Understanding and Detecting Supporting Arguments of Diverse Types , 2017 .

[44]  Benno Stein,et al.  Unit Segmentation of Argumentative Texts , 2017, ArgMining@EMNLP.

[45]  Iryna Gurevych,et al.  Parsing Argumentation Structures in Persuasive Essays , 2016, CL.

[46]  Benno Stein,et al.  Building an Argument Search Engine for the Web , 2017, ArgMining@EMNLP.

[47]  Jiliang Tang,et al.  Multi-Source Multi-Class Fake News Detection , 2018, COLING.

[48]  Serena Villata,et al.  Five Years of Argument Mining: a Data-driven Analysis , 2018, IJCAI.

[49]  Noam Slonim,et al.  Towards an argumentative content search engine using weak supervision , 2018, COLING.

[50]  Iryna Gurevych,et al.  A Retrospective Analysis of the Fake News Challenge Stance-Detection Task , 2018, COLING.

[51]  Dan Roth,et al.  TwoWingOS: A Two-Wing Optimization Strategy for Evidential Claim Verification , 2018, EMNLP.

[52]  Dan Roth,et al.  Looking Beyond the Surface: A Challenge Set for Reading Comprehension over Multiple Sentences , 2018, NAACL.

[53]  Andreas Vlachos,et al.  FEVER: a Large-scale Dataset for Fact Extraction and VERification , 2018, NAACL.

[54]  Smaranda Muresan,et al.  Where is Your Evidence: Improving Fact-checking by Justification Modeling , 2018 .

[55]  Peter Clark,et al.  SciTaiL: A Textual Entailment Dataset from Science Question Answering , 2018, AAAI.

[56]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.