Neural Architectures for Biological Inter-Sentence Relation Extraction

We introduce a family of deep-learning architectures for inter-sentence relation extraction, i.e., relations where the participants are not necessarily in the same sentence. We apply these architectures to an important use case in the biomedical domain: assigning biological context to biochemical events. In this work, biological context is defined as the type of biological system within which the biochemical event is observed. The neural architectures encode and aggregate multiple occurrences of the same candidate context mentions to determine whether it is the correct context for a particular event mention. We propose two broad types of architectures: the first type aggregates multiple instances that correspond to the same candidate context with respect to event mention before emitting a classification; the second type independently classifies each instance and uses the results to vote for the final class, akin to an ensemble approach. Our experiments show that the proposed neural classifiers are competitive and some achieve better performance than previous state of the art traditional machine learning methods without the need for feature engineering. Our analysis shows that the neural methods particularly improve precision compared to traditional machine learning classifiers and also demonstrates how the difficulty of inter-sentence relation extraction increases as the distance between the event and context mentions increase.

[1]  M. Wang,et al.  An Unsupervised Text Mining Method for Relation Extraction from Biomedical Literature , 2014, PloS one.

[2]  Daniel Marcu,et al.  Biomedical Event Extraction using Abstract Meaning Representation , 2017, BioNLP.

[3]  Andrew McCallum,et al.  Model Combination for Event Extraction in BioNLP 2011 , 2011, BioNLP@ACL.

[4]  Ming-Wei Chang,et al.  BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.

[5]  Jari Björne,et al.  Biomedical Event Extraction Using Convolutional Neural Networks and Dependency Parsing , 2018, BioNLP.

[6]  Michael Gamon,et al.  MSR-NLP Entry in BioNLP Shared Task 2011 , 2011, BioNLP@ACL.

[7]  Omer Levy,et al.  RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.

[8]  C. Der,et al.  Distinct requirements for Ras oncogenesis in human versus mouse cells. , 2002, Genes & development.

[9]  Danushka Bollegala,et al.  A Dataset for Inter-Sentence Relation Extraction using Distant Supervision , 2018, LREC.

[10]  Clayton T. Morrison,et al.  Large-scale automated machine reading discovers new cancer-driving mechanisms , 2018, Database J. Biol. Databases Curation.

[11]  Han Fang,et al.  Linformer: Self-Attention with Linear Complexity , 2020, ArXiv.

[12]  Deyu Zhou,et al.  Biomedical Relation Extraction: From Binary to Complex , 2014, Comput. Math. Methods Medicine.

[13]  Farzaneh Sarafraz,et al.  Finding conflicting statements in the biomedical literature , 2012 .

[14]  Ralf Zimmer,et al.  RelEx - Relation extraction using dependency parse trees , 2007, Bioinform..

[15]  Halil Kilicoglu,et al.  Adapting a General Semantic Interpretation Approach to Biological Event Extraction , 2011, BioNLP@ACL.

[16]  Goran Nenadic,et al.  An Exploration of Mining Gene Expression Mentions and Their Anatomical Locations from Biomedical Text , 2010, BioNLP@ACL.

[17]  Victor Cellarius,et al.  Reach , 2010, Canadian Medical Association Journal.

[18]  Mark Stevenson,et al.  Extracting Relations Within and Across Sentences , 2011, RANLP.

[19]  Doug Downey,et al.  Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.

[20]  Paul R Cohen,et al.  DARPA's Big Mechanism program , 2015, Physical biology.

[21]  Jaewoo Kang,et al.  BioBERT: a pre-trained biomedical language representation model for biomedical text mining , 2019, Bioinform..

[22]  Lukasz Kaiser,et al.  Rethinking Attention with Performers , 2020, ArXiv.

[23]  Alfonso Valencia,et al.  Overview of BioCreAtIvE: critical assessment of information extraction for biology , 2005, BMC Bioinformatics.

[24]  Sophia Ananiadou,et al.  Inter-sentence Relation Extraction with Document-level Graph Convolutional Neural Network , 2019, ACL.

[25]  Enrique Noriega-Atala,et al.  Extracting Inter-Sentence Relations for Associating Biological Context with Events in Biomedical Texts , 2020, IEEE/ACM Transactions on Computational Biology and Bioinformatics.

[26]  Maosong Sun,et al.  DocRED: A Large-Scale Document-Level Relation Extraction Dataset , 2019, ACL.

[27]  Wei-Hung Weng,et al.  Publicly Available Clinical BERT Embeddings , 2019, Proceedings of the 2nd Clinical Natural Language Processing Workshop.

[28]  Hoifung Poon,et al.  Distant Supervision for Cancer Pathway Extraction from Text , 2014, Pacific Symposium on Biocomputing.

[29]  Yi Tay,et al.  Synthesizer: Rethinking Self-Attention for Transformer Models , 2020, ICML.

[30]  Sophia Ananiadou,et al.  DeepEventMine: end-to-end neural nested event extraction from biomedical texts , 2020, Bioinform..

[31]  Oren Etzioni,et al.  Open Information Extraction from the Web , 2007, CACM.

[32]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[33]  Jari Björne,et al.  Generalizing Biomedical Event Extraction , 2011, BioNLP@ACL.

[34]  Nguyen Bach,et al.  A Review of Relation Extraction , 2007 .

[35]  Arman Cohan,et al.  Longformer: The Long-Document Transformer , 2020, ArXiv.

[36]  Peng Chen,et al.  PermuteFormer: Efficient Relative Position Encoding for Long Sequences , 2021, EMNLP.

[37]  Clayton T. Morrison,et al.  Large-scale Automated Reading with Reach Discovers New Cancer Driving Mechanisms , 2017 .