From What to Why: Improving Relation Extraction with Rationale Graph

Which type of information affects the existing neural relation extraction (RE) models to make correct decisions is an important question. In this paper, we observe that entity type and trigger are the most indicative information for RE in each instance. Moreover, these indicative clues are always constrained to co-occur with specific relations at the corpus level. Motivated by this, we propose a novel RAtionale Graph (RAG) to organize such co-occurrence constraints among entity types, triggers and relations in a holistic graph view. By introducing two subtasks of entity type prediction and trigger labeling, we build the connection between each instance and RAG, and then leverage relevant global co-occurrence knowledge stored in the graph to improve the performance of neural RE models. Extensive experimental results indicate that our method outperforms strong baselines significantly and achieves state-ofthe-art performance on the document-level and sentence-level RE benchmarks.

[1]  Chiranjib Bhattacharyya,et al.  RESIDE: Improving Distantly-Supervised Neural Relation Extraction using Side Information , 2018, EMNLP.

[2]  Omer Levy,et al.  SpanBERT: Improving Pre-training by Representing and Predicting Spans , 2019, TACL.

[3]  Jun Zhao,et al.  Relation Classification via Convolutional Deep Neural Network , 2014, COLING.

[4]  Christopher D. Manning,et al.  Graph Convolution over Pruned Dependency Trees Improves Relation Extraction , 2018, EMNLP.

[5]  Yann LeCun,et al.  Dimensionality Reduction by Learning an Invariant Mapping , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[6]  Xifeng Yan,et al.  Global Relation Embedding for Relation Extraction , 2018, NAACL-HLT.

[7]  Jiapeng Zhao,et al.  Distilling Knowledge from Well-Informed Soft Labels for Neural Relation Extraction , 2020, AAAI.

[8]  Raymond J. Mooney,et al.  Relational Learning of Pattern-Match Rules for Information Extraction , 1999, CoNLL.

[9]  Chia-Hui Chang,et al.  IEPAD: information extraction based on pattern discovery , 2001, WWW '01.

[10]  J. Tenenbaum,et al.  Special issue on “Probabilistic models of cognition , 2022 .

[11]  Zhiyuan Liu,et al.  Neural Relation Extraction with Selective Attention over Instances , 2016, ACL.

[12]  Xiao Huang,et al.  TriggerNER: Learning with Entity Triggers as Explanations for Named Entity Recognition , 2020, ACL.

[13]  Claire Cardie,et al.  Dialogue-Based Relation Extraction , 2020, ACL.

[14]  Ming Gao,et al.  Improving Neural Relation Extraction with Implicit Mutual Relations , 2019, 2020 IEEE 36th International Conference on Data Engineering (ICDE).

[15]  Richong Zhang,et al.  Relation Extraction with Convolutional Network over Learnable Syntax-Transport Graph , 2020, AAAI.

[16]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[17]  Wei Lu,et al.  Attention Guided Graph Convolutional Networks for Relation Extraction , 2019, ACL.

[18]  Maosong Sun,et al.  ERNIE: Enhanced Language Representation with Informative Entities , 2019, ACL.

[19]  Yang Liu,et al.  Exploring Fine-grained Entity Type Constraints for Distantly Supervised Relation Extraction , 2014, COLING.

[20]  Bin Wang,et al.  Beyond Word Attention: Using Segment Attention in Neural Relation Extraction , 2019, IJCAI.

[21]  Jeffrey Ling,et al.  Matching the Blanks: Distributional Similarity for Relation Learning , 2019, ACL.

[22]  Jian Su,et al.  Exploring Various Knowledge in Relation Extraction , 2005, ACL.

[23]  Philip S. Yu,et al.  SelfORE: Self-supervised Relational Feature Learning for Open Relation Extraction , 2020, EMNLP.

[24]  Max Welling,et al.  Modeling Relational Data with Graph Convolutional Networks , 2017, ESWC.

[25]  Nanda Kambhatla,et al.  Combining Lexical, Syntactic, and Semantic Features with Maximum Entropy Models for Information Extraction , 2004, ACL.

[26]  Huan Sun,et al.  Rationalizing Medical Relation Prediction from Corpus-level Statistics , 2020, ACL.

[27]  Aleksandra Gabryszak,et al.  TACRED Revisited: A Thorough Evaluation of the TACRED Relation Extraction Task , 2020, ACL.

[28]  Maosong Sun,et al.  Learning from Context or Names? An Empirical Study on Neural Relation Extraction , 2020, EMNLP.

[29]  Danqi Chen,et al.  Position-aware Attention and Supervised Data Improve Slot Filling , 2017, EMNLP.