Meta-Learning for Neural Relation Classification with Distant Supervision

Distant supervision provides a means to create a large number of weakly labeled data at low cost for relation classification. However, the resulting labeled instances are very noisy, containing data with wrong labels. Many approaches have been proposed to select a subset of reliable instances for neural model training, but they still suffer from noisy labeling problem or underutilization of the weakly-labeled data. To better select more reliable training instances, we introduce a small amount of manually labeled data as reference to guide the selection process. In this paper, we propose a meta-learning based approach, which learns to reweight noisy training data under the guidance of reference data. As the clean reference data is usually very small, we propose to augment it by dynamically distilling the most reliable elite instances from the noisy data. Experiments on several datasets demonstrate that the reference data can effectively guide the selection of training data, and our augmented approach consistently improves the performance of relation classification comparing to the existing state-of-the-art methods.

[1]  Jascha Sohl-Dickstein,et al.  Meta-Learning Update Rules for Unsupervised Representation Learning , 2018, ICLR.

[2]  Daniel Jurafsky,et al.  Distant supervision for relation extraction without labeled data , 2009, ACL.

[3]  Ebrahim Bagheri,et al.  Open Information Extraction , 2016, Encycl. Semantic Comput. Robotic Intell..

[4]  Shujian Huang,et al.  Exploiting Noisy Data in Distant Supervision Relation Classification , 2019, NAACL.

[5]  Bin Yang,et al.  Learning to Reweight Examples for Robust Deep Learning , 2018, ICML.

[6]  Zhiyuan Liu,et al.  Denoising Distant Supervision for Relation Extraction via Instance-Level Adversarial Training , 2018, ArXiv.

[7]  Jun Zhao,et al.  Distant Supervision for Relation Extraction via Piecewise Convolutional Neural Networks , 2015, EMNLP.

[8]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[9]  William Yang Wang,et al.  Robust Distant Supervision Relation Extraction via Deep Reinforcement Learning , 2018, ACL.

[10]  Liyuan Liu,et al.  Looking Beyond Label Noise: Shifted Label Distribution Matters in Distantly Supervised Relation Extraction , 2019, EMNLP.

[11]  Sebastian Thrun,et al.  Learning to Learn: Introduction and Overview , 1998, Learning to Learn.

[12]  Ralph Grishman,et al.  Infusion of Labeled Data into Distant Supervision for Relation Extraction , 2014, ACL.

[13]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[14]  Wei Shi,et al.  Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification , 2016, ACL.

[15]  Daniel S. Weld,et al.  Open Information Extraction Using Wikipedia , 2010, ACL.

[16]  Waleed Ammar,et al.  Combining Distant and Direct Supervision for Neural Relation Extraction , 2019, NAACL-HLT.

[17]  Zhifang Sui,et al.  A Soft-label Method for Noise-tolerant Distantly Supervised Relation Extraction , 2017, EMNLP.

[18]  Hai Zhao,et al.  GAN Driven Semi-distant Supervision for Relation Extraction , 2019, NAACL.

[19]  David Bamman,et al.  Adversarial Training for Relation Extraction , 2017, EMNLP.

[20]  Jing Liu,et al.  Revisiting Distant Supervision for Relation Extraction , 2018, LREC.

[21]  Danqi Chen,et al.  Position-aware Attention and Supervised Data Improve Slot Filling , 2017, EMNLP.

[22]  Percy Liang,et al.  Understanding Black-box Predictions via Influence Functions , 2017, ICML.

[23]  Jun Zhao,et al.  Distant Supervision for Relation Extraction with Sentence-Level Attention and Entity Descriptions , 2017, AAAI.

[24]  Luke S. Zettlemoyer,et al.  Knowledge-Based Weak Supervision for Information Extraction of Overlapping Relations , 2011, ACL.

[25]  Daniel S. Weld,et al.  Autonomously semantifying wikipedia , 2007, CIKM '07.

[26]  Jun Zhao,et al.  Large Scaled Relation Extraction With Reinforcement Learning , 2018, AAAI.

[27]  Zhiyuan Liu,et al.  Neural Relation Extraction with Selective Attention over Instances , 2016, ACL.

[28]  Xinyan Xiao,et al.  ARNOR: Attention Regularization based Noise Reduction for Distant Supervision Relation Classification , 2019, ACL.

[29]  Qi Xie,et al.  Meta-Weight-Net: Learning an Explicit Mapping For Sample Weighting , 2019, NeurIPS.

[30]  William Yang Wang,et al.  DSGAN: Generative Adversarial Training for Distant Supervision Relation Extraction , 2018, ACL.

[31]  Joshua Achiam,et al.  On First-Order Meta-Learning Algorithms , 2018, ArXiv.

[32]  Tao Shen,et al.  Self-Attention Enhanced Selective Gate with Entity-Aware Embedding for Distantly Supervised Relation Extraction , 2019, AAAI.

[33]  Andreas Vlachos,et al.  Meta-Learning Improves Lifelong Relation Extraction , 2019, RepL4NLP@ACL.

[34]  Daniel S. Weld,et al.  Fine-Grained Entity Recognition , 2012, AAAI.

[35]  Dantong Ouyang,et al.  A Fine-grained and Noise-aware Method for Neural Relation Extraction , 2019, CIKM.

[36]  Jun Zhao,et al.  Relation Classification via Convolutional Deep Neural Network , 2014, COLING.

[37]  Andreas Vlachos,et al.  Model-Agnostic Meta-Learning for Relation Classification with Limited Supervision , 2019, ACL.

[38]  Stephanie M. Strassel,et al.  Linguistic Resources for 2013 Knowledge Base Population Evaluations , 2012 .

[39]  Xuchen Yao,et al.  Information Extraction over Structured Data: Question Answering with Freebase , 2014, ACL.

[40]  Li Zhao,et al.  Reinforcement Learning for Relation Classification From Noisy Data , 2018, AAAI.

[41]  Zhiyuan Liu,et al.  Relation Classification via Multi-Level Attention CNNs , 2016, ACL.

[42]  Ramesh Nallapati,et al.  Multi-instance Multi-label Learning for Relation Extraction , 2012, EMNLP.

[43]  Andrew McCallum,et al.  Modeling Relations and Their Mentions without Labeled Text , 2010, ECML/PKDD.

[44]  Zhen-Hua Ling,et al.  Distant Supervision Relation Extraction with Intra-Bag and Inter-Bag Attentions , 2019, NAACL.