A Novel Approach for Relation Extraction with Few Labeled Data

Lack of large scale training data is a challenge for conventional supervised relation extraction approach. Although distant supervision has been proposed to address this issue, it suffers from massive noise and the trained model cannot be applied to unseen relations. We present a novel approach for relation extraction which uses the relation definition as a guide and only needs a hundred of high-quality mention examples for training model. In detail, we classify the candidate mention of a specific relation by judging whether the mention is in conformity with the relation’s definition through measuring the semantic relevance between the definition and the mention. Our approach is insensitive to class-imbalance problem. And the trained model can be directly applied to classify mentions of newly defined relation without labeling new training data. Experimental results demonstrate that our approach achieves competitive performance and can be incorporated with existing approaches to boost performance.

[1]  Guodong Zhou,et al.  Tree Kernel-Based Relation Extraction with Context-Sensitive Structured Parse Tree Information , 2007, EMNLP.

[2]  Andrew McCallum,et al.  Modeling Relations and Their Mentions without Labeled Text , 2010, ECML/PKDD.

[3]  Ralph Grishman,et al.  Infusion of Labeled Data into Distant Supervision for Relation Extraction , 2014, ACL.

[4]  Dmitry Zelenko,et al.  Kernel methods for relation extraction , 2003 .

[5]  Tiziano Flati,et al.  Two Is Bigger (and Better) Than One: the Wikipedia Bitaxonomy Project , 2014, ACL.

[6]  Jian Su,et al.  Modeling Commonality among Related Classes in Relation Extraction , 2006, ACL.

[7]  Nanda Kambhatla,et al.  Combining Lexical, Syntactic, and Semantic Features with Maximum Entropy Models for Information Extraction , 2004, ACL.

[8]  Christopher D. Manning,et al.  Combining Distant and Partial Supervision for Relation Extraction , 2014, EMNLP.

[9]  Simone Paolo Ponzetto,et al.  BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network , 2012, Artif. Intell..

[10]  Ramesh Nallapati,et al.  Multi-instance Multi-label Learning for Relation Extraction , 2012, EMNLP.

[11]  Roberto Navigli,et al.  Entity Linking meets Word Sense Disambiguation: a Unified Approach , 2014, TACL.

[12]  Eneko Agirre,et al.  Improving distant supervision using inference learning , 2015, ACL.

[13]  Ralph Grishman,et al.  Distant Supervision for Relation Extraction with an Incomplete Knowledge Base , 2013, NAACL.

[14]  Razvan C. Bunescu,et al.  A Shortest Path Dependency Kernel for Relation Extraction , 2005, HLT.

[15]  Luke S. Zettlemoyer,et al.  Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations , 2011, ACL.

[16]  Dietrich Klakow,et al.  Combining Generative and Discriminative Model Scores for Distant Supervision , 2013, EMNLP.

[17]  Jian Su,et al.  Exploring Various Knowledge in Relation Extraction , 2005, ACL.

[18]  Hiroshi Nakagawa,et al.  Reducing Wrong Labels in Distant Supervision for Relation Extraction , 2012, ACL.

[19]  Jian Su,et al.  A Composite Kernel to Extract Relations between Entities with Both Flat and Structured Features , 2006, ACL.

[20]  Daniel Jurafsky,et al.  Distant supervision for relation extraction without labeled data , 2009, ACL.

[21]  John B. Lowe,et al.  The Berkeley FrameNet Project , 1998, ACL.