Combining Distant and Partial Supervision for Relation Extraction

Broad-coverage relation extraction either requires expensive supervised training data, or suffers from drawbacks inherent to distant supervision. We present an approach for providing partial supervision to a distantly supervised relation extractor using a small number of carefully selected examples. We compare against established active learning criteria and propose a novel criterion to sample examples which are both uncertain and representative. In this way, we combine the benefits of fine-grained supervision for difficult examples with the coverage of a large distantly supervised corpus. Our approach gives a substantial increase of 3.9% endto-end F1 on the 2013 KBP Slot Filling evaluation, yielding a net F1 of 37.7%.

[1]  Aidan Finn,et al.  Active Learning Selection Strategies for Information Extraction , 2003 .

[2]  Christopher Ré,et al.  Big Data versus the Crowd: Looking for Relationships in All the Right Places , 2012, ACL.

[3]  Angel X. Chang,et al.  Stanford's 2013 KBP System , 2013, TAC.

[4]  Mark Craven,et al.  Multiple-Instance Active Learning , 2007, NIPS.

[5]  Andrew McCallum,et al.  Employing EM and Pool-Based Active Learning for Text Classification , 1998, ICML.

[6]  Andrew McCallum,et al.  Relation Extraction with Matrix Factorization and Universal Schemas , 2013, NAACL.

[7]  Daniel S. Weld,et al.  Autonomously semantifying wikipedia , 2007, CIKM '07.

[8]  Noah A. Smith,et al.  Novel estimation methods for unsupervised discovery of latent structure in natural language text , 2007 .

[9]  Dietrich Klakow,et al.  Feature-based models for improving the quality of noisy training data for relation extraction , 2013, CIKM.

[10]  Le Zhao,et al.  Filling Knowledge Base Gaps for Distant Supervision of Relation Extraction , 2013, ACL.

[11]  Mark A. Przybocki,et al.  The Automatic Content Extraction (ACE) Program – Tasks, Data, and Evaluation , 2004, LREC.

[12]  H. Sebastian Seung,et al.  Information, Prediction, and Query by Committee , 1992, NIPS.

[13]  Mark Craven,et al.  Constructing Biological Knowledge Bases by Extracting Information from Text Sources , 1999, ISMB.

[14]  Luke S. Zettlemoyer,et al.  Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations , 2011, ACL.

[15]  Valentin I. Spitkovsky,et al.  Stanford's Distantly-Supervised Slot-Filling System , 2011, TAC.

[16]  Andrew McCallum,et al.  Modeling Relations and Their Mentions without Labeled Text , 2010, ECML/PKDD.

[17]  Daniel Jurafsky,et al.  Distant supervision for relation extraction without labeled data , 2009, ACL.

[18]  Ralph Grishman,et al.  New York University 2012 System for KBP Slot Filling , 2012, TAC.

[19]  Ralph Grishman,et al.  Distant Supervision for Relation Extraction with an Incomplete Knowledge Base , 2013, NAACL.

[20]  Foster J. Provost,et al.  Active Sampling for Class Probability Estimation and Ranking , 2004, Machine Learning.

[21]  Ralph Grishman,et al.  New York University KBP 2010 Slot-Filling System , 2010, TAC.

[22]  Jian Su,et al.  Exploring Various Knowledge in Relation Extraction , 2005, ACL.

[23]  William A. Gale,et al.  A sequential algorithm for training text classifiers , 1994, SIGIR '94.

[24]  Ralph Grishman,et al.  An Efficient Active Learning Framework for New Relation Types , 2013, IJCNLP.

[25]  Ramesh Nallapati,et al.  Multi-instance Multi-label Learning for Relation Extraction , 2012, EMNLP.

[26]  Ralph Grishman,et al.  Infusion of Labeled Data into Distant Supervision for Relation Extraction , 2014, ACL.

[27]  Stephen Glenn Soderland,et al.  Learning text analysis rules for domain-specific natural language processing , 1996 .

[28]  H. Sebastian Seung,et al.  Selective Sampling Using the Query by Committee Algorithm , 1997, Machine Learning.

[29]  Mihai Surdeanu,et al.  Robust Information Extraction with Perceptrons , 2007 .

[30]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .