Efficient Document-level Event Extraction via Pseudo-Trigger-aware Pruned Complete Graph

There are two main challenges in document-level event extraction: 1) argument entities are scattered in different sentences, and 2) event triggers are often not available. To address these challenges, most previous studies mainly focus on building argument chains in an autoregressive way, which is inefficient in both training and inference. In contrast to the previous studies, we propose a fast and lightweight model named as PTPCG. We design a non-autoregressive decoding algorithm to perform event argument combination extraction on pruned complete graphs, which are constructed under the guidance of the automatically selected pseudo triggers. Compared to the previous systems, our system achieves competitive results with lower resource consumption, taking only 3.6% GPU time (pfs-days) for training and up to 8.5 times faster for inference. Besides, our approach shows superior compatibility for the datasets with (or without) triggers and the pseudo triggers can be the supplements for annotated triggers to make further improvements.

[1]  Ralph Grishman,et al.  Joint Event Extraction via Recurrent Neural Networks , 2016, NAACL.

[2]  Hannaneh Hajishirzi,et al.  Entity, Relation, and Event Extraction with Contextualized Span Representations , 2019, EMNLP.

[3]  C. Bron,et al.  Algorithm 457: finding all cliques of an undirected graph , 1973 .

[4]  Runxin Xu,et al.  Document-level Event Extraction via Heterogeneous Graph-based Interaction Model with a Tracker , 2021, ACL.

[5]  Guillaume Lample,et al.  Neural Architectures for Named Entity Recognition , 2016, NAACL.

[6]  Yang Xiao,et al.  DCFEE: A Document-level Chinese Financial Event Extraction System based on Automatically Labeled Training Data , 2018, ACL.

[7]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[8]  Daniel Jurafsky,et al.  Distant supervision for relation extraction without labeled data , 2009, ACL.

[9]  Samy Bengio,et al.  Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.

[10]  Xiang Zhang,et al.  Automatically Labeled Data Generation for Large Scale Event Extraction , 2017, ACL.

[11]  Jun Zhao,et al.  Event Extraction via Dynamic Multi-Pooling Convolutional Neural Networks , 2015, ACL.

[12]  Nathan Schneider,et al.  Association for Computational Linguistics: Human Language Technologies , 2011 .

[13]  Wei Xu,et al.  Doc2EDAG: An End-to-End Document-level Framework for Chinese Financial Event Extraction , 2019, EMNLP.

[14]  Xiao Liu,et al.  Jointly Multiple Events Extraction via Attention-based Graph Information Aggregation , 2018, EMNLP.

[15]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[16]  Hang Yang,et al.  Document-level Event Extraction via Parallel Prediction Networks , 2021, ACL.

[17]  Ying Lin,et al.  A Joint Neural Model for Information Extraction with Global Features , 2020, ACL.