Dependency parsing of biomedical text with BERT
暂无分享,去创建一个
[1] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[2] Pablo Gamallo,et al. Dependency-Based Open Information Extraction , 2012 .
[3] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[4] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[5] Andrew McCallum,et al. Fast and Robust Joint Models for Biomedical Event Extraction , 2011, EMNLP.
[6] Eugene Charniak,et al. A Maximum-Entropy-Inspired Parser , 2000, ANLP.
[7] Koby Crammer,et al. Online Large-Margin Training of Dependency Parsers , 2005, ACL.
[8] Daniel Kondratyuk,et al. 75 Languages, 1 Model: Parsing Universal Dependencies Universally , 2019, EMNLP.
[9] Tapio Salakoski,et al. Distributional Semantics Resources for Biomedical Text Processing , 2013 .
[10] Timothy Dozat,et al. Stanford’s Graph-based Neural Dependency Parser at the CoNLL 2017 Shared Task , 2017, CoNLL.
[11] Slav Petrov,et al. Globally Normalized Transition-Based Neural Networks , 2016, ACL.
[12] Sampo Pyysalo,et al. Neural Dependency Parsing of Biomedical Text: TurkuNLP entry in the CRAFT Structural Annotation Task , 2019, EMNLP.
[13] Yijia Liu,et al. Towards Better UD Parsing: Deep Contextualized Word Embeddings, Ensemble, and Treebank Concatenation , 2018, CoNLL.
[14] Iz Beltagy,et al. SciBERT: A Pretrained Language Model for Scientific Text , 2019, EMNLP.
[15] Philipp Koehn,et al. Synthesis Lectures on Human Language Technologies , 2016 .
[16] Sampo Pyysalo,et al. Universal Dependencies v2: An Evergrowing Multilingual Treebank Collection , 2020, LREC.
[17] Daniel Zeman,et al. Proceedings of the CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies , 2017, CoNLL Shared Task.
[18] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[19] Tapio Salakoski,et al. Turku Neural Parser Pipeline: An End-to-End System for the CoNLL 2018 Shared Task , 2018, CoNLL.
[20] Sampo Pyysalo,et al. Universal Dependencies v1: A Multilingual Treebank Collection , 2016, LREC.
[21] Lawrence Hunter,et al. CRAFT Shared Tasks 2019 Overview — Integrated Structure, Semantics, and Coreference , 2019, EMNLP.
[22] Dan Klein,et al. Accurate Unlexicalized Parsing , 2003, ACL.
[23] Gosse Bouma,et al. Overview of the IWPT 2020 Shared Task on Parsing into Enhanced Universal Dependencies , 2020, IWPT.
[24] Beatrice Santorini,et al. The Penn Treebank: An Overview , 2003 .
[25] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[26] Joakim Nivre,et al. Universal Stanford dependencies: A cross-linguistic typology , 2014, LREC.
[27] Jari Björne,et al. Deep Learning with Minimal Training Data: TurkuNLP Entry in the BioNLP Shared Task 2016 , 2016, BioNLP.
[28] Joachim Bingel,et al. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics , 2016 .
[29] Noah A. Smith,et al. Dependency Parsing , 2009, Encyclopedia of Artificial Intelligence.
[30] Kenji Sagae,et al. Dynamic Programming for Linear-Time Incremental Parsing , 2010, ACL.
[31] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[32] Daniel M. Bikel,et al. A Distributional Analysis of a Lexicalized Statistical Parsing Model , 2004, EMNLP.
[33] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[34] Joakim Nivre,et al. Transition-based Dependency Parsing with Rich Non-local Features , 2011, ACL.
[35] Wiebke Wagner,et al. Steven Bird, Ewan Klein and Edward Loper: Natural Language Processing with Python, Analyzing Text with the Natural Language Toolkit , 2010, Lang. Resour. Evaluation.
[36] Tapio Salakoski,et al. Universal Lemmatizer: A sequence-to-sequence model for lemmatizing Universal Dependencies treebanks , 2019, Natural Language Engineering.
[37] Nizar Habash,et al. CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies , 2017, CoNLL.
[38] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[39] Sampo Pyysalo,et al. How to Train good Word Embeddings for Biomedical NLP , 2016, BioNLP@ACL.
[40] K. Bretonnel Cohen,et al. Concept annotation in the CRAFT corpus , 2012, BMC Bioinformatics.
[41] Joakim Nivre,et al. Labeled Pseudo-Projective Dependency Parsing with Support Vector Machines , 2006, CoNLL.
[42] Zhiyong Lu,et al. Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets , 2019, BioNLP@ACL.
[43] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[44] Milan Straka,et al. Tokenizing, POS Tagging, Lemmatizing and Parsing UD 2.0 with UDPipe , 2017, CoNLL.
[45] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[46] Christopher D. Manning,et al. Stanford typed dependencies manual , 2010 .
[47] Noah A. Smith,et al. Transition-Based Dependency Parsing with Stack Long Short-Term Memory , 2015, ACL.
[48] Sampo Pyysalo,et al. Turku Enhanced Parser Pipeline: From Raw Text to Enhanced Graphs in the IWPT 2020 Shared Task , 2020, IWPT.
[49] Razvan C. Bunescu,et al. A Shortest Path Dependency Kernel for Relation Extraction , 2005, HLT.
[50] Martin Potthast,et al. CoNLL 2018 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies , 2018, CoNLL.
[51] Jari Björne,et al. Complex event extraction at PubMed scale , 2010, Bioinform..
[52] Danqi Chen,et al. A Fast and Accurate Dependency Parser using Neural Networks , 2014, EMNLP.
[53] K. Bretonnel Cohen,et al. A corpus of full-text journal articles is a robust evaluation tool for revealing differences in performance of biomedical natural language processing tools , 2012, BMC Bioinformatics.
[54] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[55] Jaewoo Kang,et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining , 2019, Bioinform..
[56] Eugene Charniak,et al. Coarse-to-Fine n-Best Parsing and MaxEnt Discriminative Reranking , 2005, ACL.