Behind the Scenes: An Exploration of Trigger Biases Problem in Few-Shot Event Classification

Few-Shot Event Classification (FSEC) aims at developing a model for event prediction, which can generalize to new event types with a limited number of annotated data. Existing FSEC studies have achieved high accuracy on different benchmarks. However, we find they suffer from trigger biases that signify the statistical homogeneity between some trigger words and target event types, which we summarize as trigger overlapping and trigger separability. The biases can result in context-bypassing problem, i.e., correct classifications can be gained by looking at only the trigger words while ignoring the entire context. Therefore, existing models can be weak in generalizing to unseen data in real scenarios. To further uncover the trigger biases and assess the generalization ability of the models, we propose two new sampling methods, Trigger-Uniform Sampling (TUS) and COnfusion Sampling (COS), for the meta tasks construction during evaluation. Besides, to cope with the context-bypassing problem in FSEC models, we introduce adversarial training and trigger reconstruction techniques. Experiments show these techniques help not only improve the performance, but also enhance the generalization ability of models.

[1]  Omer Levy,et al.  Annotation Artifacts in Natural Language Inference Data , 2018, NAACL.

[2]  Shuang Zeng,et al.  Double Graph Based Reasoning for Document-level Relation Extraction , 2020, EMNLP.

[3]  Jian Liu,et al.  Exploiting the Ground-Truth: An Adversarial Imitation Based Knowledge Distillation Approach for Event Detection , 2019, AAAI.

[4]  Franck Dernoncourt,et al.  Exploiting the Matching Information in the Support Set for Few Shot Event Classification , 2020, PAKDD.

[5]  Xing Xie,et al.  REET: Joint Relation Extraction and Entity Typing via Multi-task Learning , 2019, NLPCC.

[6]  Xiaoli Z. Fern,et al.  Event Nugget Detection with Forward-Backward Recurrent Neural Networks , 2016, ACL.

[7]  Lifu Tu,et al.  Pay Attention to the Ending:Strong Neural Baselines for the ROC Story Cloze Task , 2017, ACL.

[8]  Xiaocheng Feng,et al.  A language-independent neural network for event detection , 2016, Science China Information Sciences.

[9]  R. Thomas McCoy,et al.  Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference , 2019, ACL.

[10]  Omer Levy,et al.  Do Supervised Distributional Methods Really Learn Lexical Inference Relations? , 2015, NAACL.

[11]  Yash Goyal,et al.  Making the V in VQA Matter: Elevating the Role of Image Understanding in Visual Question Answering , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Sameer Singh,et al.  Compositional Questions Do Not Necessitate Multi-hop Reasoning , 2019, ACL.

[13]  Xin Zheng,et al.  An Empirical Study on Model-agnostic Debiasing Strategies for Robust Natural Language Inference , 2020, CONLL.

[14]  Ralph Grishman,et al.  Graph Convolutional Networks With Argument-Aware Pooling for Event Detection , 2018, AAAI.

[15]  Jie Zhou,et al.  MAVEN: A Massive General Domain Event Detection Dataset , 2020, EMNLP.

[16]  Huajun Chen,et al.  Meta-Learning with Dynamic-Memory-Based Prototypical Network for Few-Shot Event Detection , 2020, WSDM.

[17]  Xin Zheng,et al.  HypoNLI: Exploring the Artificial Patterns of Hypothesis-only Bias in Natural Language Inference , 2020, LREC.

[18]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[19]  Fangzhao Wu,et al.  REKA: Relation Extraction with Knowledge-Aware Attention , 2019, CCKS.

[20]  Franck Dernoncourt,et al.  Extensively Matching for Few-shot Learning Event Detection , 2020, NUSE.

[21]  Yonatan Belinkov,et al.  Don’t Take the Premise for Granted: Mitigating Artifacts in Natural Language Inference , 2019, ACL.

[22]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[23]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[24]  Yejin Choi,et al.  The Effect of Different Writing Tasks on Linguistic Style: A Case Study of the ROC Story Cloze Task , 2017, CoNLL.

[25]  Richard S. Zemel,et al.  Prototypical Networks for Few-shot Learning , 2017, NIPS.

[26]  Bowen Yu,et al.  Few-Shot Event Detection with Prototypical Amortized Conditional Random Field , 2021, FINDINGS.

[27]  Jun Zhao,et al.  Exploiting Argument Information to Improve Event Detection via Supervised Attention Mechanisms , 2017, ACL.

[28]  Kevin Gimpel,et al.  Discriminatively-Tuned Generative Classifiers for Robust Natural Language Inference , 2020, EMNLP.

[29]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[30]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[31]  Andrew M. Dai,et al.  Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.

[32]  Yaojie Lu,et al.  Nugget Proposal Networks for Chinese Event Detection , 2018, ACL.

[33]  Shi Feng,et al.  Misleading Failures of Partial-input Baselines , 2019, ACL.

[34]  Dirk Hovy,et al.  Predictive Biases in Natural Language Processing Models: A Conceptual Framework and Overview , 2019, ACL.

[35]  Zhiyuan Liu,et al.  Hybrid Attention-Based Prototypical Networks for Noisy Few-Shot Relation Classification , 2019, AAAI.

[36]  Runxin Xu,et al.  Document-level Event Extraction via Heterogeneous Graph-based Interaction Model with a Tracker , 2021, ACL.