Towards Few-Shot Event Mention Retrieval: An Evaluation Framework and A Siamese Network Approach

Automatically analyzing events in a large amount of text is crucial for situation awareness and decision making. Previous approaches treat event extraction as “one size fits all” with an ontology defined a priori. The resulted extraction models are built just for extracting those types in the ontology. These approaches cannot be easily adapted to new event types nor new domains of interest. To accommodate personalized event-centric information needs, this paper introduces the few-shot Event Mention Retrieval (EMR) task: given a user-supplied query consisting of a handful of event mentions, return relevant event mentions found in a corpus. This formulation enables “query by example”, which drastically lowers the bar of specifying event-centric information needs. The retrieval setting also enables fuzzy search. We present an evaluation framework leveraging existing event datasets such as ACE. We also develop a Siamese Network approach, and show that it performs better than ad-hoc retrieval models in the few-shot EMR setting.

[1]  David Ahn,et al.  The stages of event extraction , 2006 .

[2]  Jamie Callan,et al.  Deeper Text Understanding for IR with Contextual Neural Language Modeling , 2019, SIGIR.

[3]  James Allan,et al.  Topic detection and tracking: event-based information organization , 2002 .

[4]  Ellen Riloff,et al.  Modeling Textual Cohesion for Event Extraction , 2012, AAAI.

[5]  Benoît Sagot,et al.  What Does BERT Learn about the Structure of Language? , 2019, ACL.

[6]  Jun'ichi Tsujii,et al.  Corpus annotation for mining biomedical events from literature , 2008, BMC Bioinformatics.

[7]  Ralph Grishman,et al.  A Two-stage Approach for Extending Event Detection to New Types via Neural Networks , 2016, Rep4NLP@ACL.

[8]  Mark A. Przybocki,et al.  The Automatic Content Extraction (ACE) Program – Tasks, Data, and Evaluation , 2004, LREC.

[9]  Douglas W. Oard,et al.  A Test Collection for Coreferent Mention Retrieval , 2018, SIGIR.

[10]  Seth Kulick,et al.  From Light to Rich ERE: Annotation of Entities, Relations, and Events , 2015, EVENTS@HLP-NAACL.

[11]  Gregory R. Koch,et al.  Siamese Neural Networks for One-Shot Image Recognition , 2015 .

[12]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[13]  Teruko Mitamura,et al.  Overview of TAC KBP 2015 Event Nugget Track , 2015, TAC.

[14]  Dipanjan Das,et al.  BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.

[15]  Omer Levy,et al.  What Does BERT Look at? An Analysis of BERT’s Attention , 2019, BlackboxNLP@ACL.

[16]  Heng Ji,et al.  Refining Event Extraction through Cross-Document Inference , 2008, ACL.

[17]  Jun Zhao,et al.  Event Extraction via Dynamic Multi-Pooling Convolutional Neural Networks , 2015, ACL.

[18]  Dan Roth,et al.  Event Detection and Co-reference with Minimal Supervision , 2016, EMNLP.

[19]  Ann Bies,et al.  Situational Awareness for Low Resource Languages: the LORELEI Situation Frame Annotation Task , 2017, SMERP@ECIR.

[20]  Dongsheng Li,et al.  Exploring Pre-trained Language Models for Event Extraction and Generation , 2019, ACL.

[21]  Ralph Grishman,et al.  Message Understanding Conference- 6: A Brief History , 1996, COLING.