SynSeq4ED: A Novel Event-Aware Text Representation Learning for Event Detection

Event detection (ED) is considered as an important task in natural language processing (NLP) which effectively supports to specify instances of multi event types which are mentioned in text. Recent models adopt advanced neural network architectures, such as long short-term memory (LSTM), graph convolutional network (GCN), etc. to capture the sequential and syntactical representations of texts for leveraging the performance of ED. However, recent neural network-based models neglect to sufficiently perverse both sequential comprehensive meanings as well as syntactical co-referencing relationships between words in the sentences. In this paper, we proposed a novel integration of GCN-based textual syntactical encoder and pre-trained BERT sequential embedding with event-aware masked language mechanism, called SynSeq4ED. In our SynSeq4ED model, we formally present a joint text embedding framework which enable to effectively learn the deep semantic representations of event triggers and arguments by introducing a combination of integrated pre-trained BERT with event-aware masked language strategy and GCN-based syntactical co-referencing text encoding mechanism. The achieved text representations by SynSeq4ED model are then used to improve the performance of multiple tasks in ED, including multiple event detection (MED), few-shot learning event detection (FSLED). Extensive experiments in benchmark datasets demonstrate the effectiveness of our proposed SynSeq4ED model in comparing with recent state-of-the-art baselines.

[1]  Zhifang Sui,et al.  Jointly Extracting Event Triggers and Arguments by Dependency-Bridge RNN and Tensor-Based Argument Interaction , 2018, AAAI.

[2]  Germán Sanchis-Trilles,et al.  Combining Embeddings of Input Data for Text Classification , 2020, Neural Processing Letters.

[3]  Jiangtao Ren,et al.  Label-Embedding Bi-directional Attentive Model for Multi-label Text Classification , 2021, Neural Process. Lett..

[4]  Xiao Liu,et al.  Jointly Multiple Events Extraction via Attention-based Graph Information Aggregation , 2018, EMNLP.

[5]  Ming Zhu,et al.  A Novel Two-stage Learning Pipeline for Deep Neural Networks , 2017, Neural Processing Letters.

[6]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[7]  Eduard H. Hovy,et al.  When Are Tree Structures Necessary for Deep Learning of Representations? , 2015, EMNLP.

[8]  Tom M. Mitchell,et al.  Joint Extraction of Events and Entities within a Document Context , 2016, NAACL.

[9]  Bang Wang,et al.  A Survey of Event Extraction From Text , 2019, IEEE Access.

[10]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[11]  Jun Zhao,et al.  Event Extraction via Dynamic Multi-Pooling Convolutional Neural Networks , 2015, ACL.

[12]  Jun Zhao,et al.  Event Extraction via Bidirectional Long Short-Term Memory Tensor Neural Networks , 2016, CCL.

[13]  Xiaoli Z. Fern,et al.  Event Nugget Detection with Forward-Backward Recurrent Neural Networks , 2016, ACL.

[14]  Ralph Grishman,et al.  Event Detection and Domain Adaptation with Convolutional Neural Networks , 2015, ACL.

[15]  Ralph Grishman,et al.  Graph Convolutional Networks With Argument-Aware Pooling for Event Detection , 2018, AAAI.

[16]  Ralph Grishman,et al.  Joint Event Extraction via Recurrent Neural Networks , 2016, NAACL.

[17]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[18]  Jun Zhao,et al.  A Probabilistic Soft Logic Based Approach to Exploiting Latent and Global Information in Event Classification , 2016, AAAI.

[19]  Ralph Grishman,et al.  Modeling Skip-Grams for Event Detection with Convolutional Neural Networks , 2016, EMNLP.

[20]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[21]  Chen Chen,et al.  Relieving the Computational Bottleneck: Joint Inference for Event Extraction with High-Dimensional Features , 2014, EMNLP.

[22]  Christopher D. Manning,et al.  Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.

[23]  Ji Wu,et al.  Tensor Graph Convolutional Networks for Text Classification , 2020, AAAI.

[24]  Ting Liu,et al.  Learning Target-Dependent Sentence Representations for Chinese Event Detection , 2018, CCIR.

[25]  Franck Dernoncourt,et al.  Extensively Matching for Few-shot Learning Event Detection , 2020, NUSE.

[26]  Tomas Mikolov,et al.  Advances in Pre-Training Distributed Word Representations , 2017, LREC.

[27]  Mihai Surdeanu,et al.  The Stanford CoreNLP Natural Language Processing Toolkit , 2014, ACL.