Pretrained Language Models for Biomedical and Clinical Tasks: Understanding and Extending the State-of-the-Art
暂无分享,去创建一个
Veselin Stoyanov | Myle Ott | Jingfei Du | Patrick Lewis | Myle Ott | Jingfei Du | Veselin Stoyanov | Patrick Lewis
[1] Zhiyong Lu,et al. Transfer Learning in Biomedical Natural Language Processing: An Evaluation of BERT and ELMo on Ten Benchmarking Datasets , 2019, BioNLP@ACL.
[2] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[3] Naveen Arivazhagan,et al. Small and Practical BERT Models for Sequence Labeling , 2019, EMNLP.
[4] Iz Beltagy,et al. SciBERT: A Pretrained Language Model for Scientific Text , 2019, EMNLP.
[5] Jaewoo Kang,et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining , 2019, Bioinform..
[6] Frank Hutter,et al. Decoupled Weight Decay Regularization , 2017, ICLR.
[7] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[8] Jingqi Wang,et al. Enhancing Clinical Concept Extraction with Contextual Embedding , 2019, J. Am. Medical Informatics Assoc..
[9] Ming-Wei Chang,et al. Well-Read Students Learn Better: On the Importance of Pre-training Compact Models , 2019 .
[10] Gary D. Bader,et al. Transfer learning for biomedical named entity recognition with neural networks , 2018, bioRxiv.
[11] Wei-Hung Weng,et al. Publicly Available Clinical BERT Embeddings , 2019, Proceedings of the 2nd Clinical Natural Language Processing Workshop.
[12] Shuying Shen,et al. 2010 i2b2/VA challenge on concepts, assertions, and relations in clinical text , 2011, J. Am. Medical Informatics Assoc..
[13] Samuel R. Bowman,et al. Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks , 2018, ArXiv.
[14] Zhiyong Lu,et al. NCBI disease corpus: A resource for disease name recognition and concept normalization , 2014, J. Biomed. Informatics.
[15] Doug Downey,et al. Construction of the Literature Graph in Semantic Scholar , 2018, NAACL.
[16] Anália Lourenço,et al. Overview of the BioCreative VI chemical-protein interaction Track , 2017 .
[17] Paloma Martínez,et al. The DDI corpus: An annotated corpus with pharmacological substances and drug-drug interactions , 2013, J. Biomed. Informatics.
[18] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[19] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[20] Yiming Yang,et al. MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices , 2020, ACL.
[21] Qingyu Chen,et al. BioWordVec, improving biomedical word embeddings with subword information and MeSH , 2019, Scientific Data.
[22] Özlem Uzuner,et al. Annotating longitudinal clinical narratives for de-identification: The 2014 i2b2/UTHealth corpus , 2015, J. Biomed. Informatics.
[23] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[24] Xiaodong Liu,et al. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing , 2020, ArXiv.
[25] Ioannis Ch. Paschalidis,et al. Clinical Concept Extraction with Contextual Word Embedding , 2018, NIPS 2018.
[26] Núria Queralt-Rosinach,et al. Extraction of relations between genes and diseases from text and large-scale data analysis: implications for translational research , 2014, BMC Bioinformatics.
[27] Qingyu Chen,et al. An Empirical Study of Multi-Task Learning on BERT for Biomedical Text Mining , 2020, BIONLP.
[28] Goran Nenadic,et al. LINNAEUS: A species name identification system for biomedical literature , 2010, BMC Bioinformatics.
[29] Thomas Wolf,et al. DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter , 2019, ArXiv.
[30] Zhiyong Lu,et al. BioCreative V CDR task corpus: a resource for chemical disease relation extraction , 2016, Database J. Biol. Databases Curation.
[31] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[32] Omer Levy,et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems , 2019, NeurIPS.
[33] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[34] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[35] Richard Tzong-Han Tsai,et al. Overview of BioCreative II gene mention recognition , 2008, Genome Biology.
[36] Balu Bhasuran,et al. Automatic extraction of gene-disease associations from literature using joint ensemble learning , 2018, PloS one.
[37] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[38] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[39] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[40] Jaewoo Kang,et al. CollaboNet: collaboration of deep neural networks for biomedical named entity recognition , 2018, BMC Bioinformatics.
[41] L. Jensen,et al. The SPECIES and ORGANISMS Resources for Fast and Accurate Identification of Taxonomic Names in Text , 2013, PloS one.
[42] Sampo Pyysalo,et al. How to Train good Word Embeddings for Biomedical NLP , 2016, BioNLP@ACL.
[43] Anna Rumshisky,et al. Evaluating temporal relations in clinical text: 2012 i2b2 Challenge , 2013, J. Am. Medical Informatics Assoc..
[44] Hongfang Liu,et al. A Comparison of Word Embeddings for the Biomedical Natural Language Processing , 2018, J. Biomed. Informatics.
[45] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[46] Anna Rumshisky,et al. Annotating temporal information in clinical narratives , 2013, J. Biomed. Informatics.
[47] Laura Inés Furlong,et al. The EU-ADR corpus: Annotated drugs, diseases, targets, and their relationships , 2012, J. Biomed. Informatics.
[48] Anna Korhonen,et al. Automatic semantic classification of scientific literature according to the hallmarks of cancer , 2016, Bioinform..
[49] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[50] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[51] Taku Kudo,et al. Subword Regularization: Improving Neural Network Translation Models with Multiple Subword Candidates , 2018, ACL.
[52] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[53] Alexey Romanov,et al. Lessons from Natural Language Inference in the Clinical Domain , 2018, EMNLP.
[54] Nigel Collier,et al. Introduction to the Bio-entity Recognition Task at JNLPBA , 2004, NLPBA/BioNLP.
[55] Jason Weston,et al. Don't Say That! Making Inconsistent Dialogue Unlikely with Unlikelihood Training , 2020, ACL.
[56] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[57] Zhiyong Lu,et al. The CHEMDNER corpus of chemicals and drugs and its annotation principles , 2015, Journal of Cheminformatics.
[58] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[59] Özlem Uzuner,et al. Automated systems for the de-identification of longitudinal clinical narratives: Overview of 2014 i2b2/UTHealth shared task Track 1 , 2015, J. Biomed. Informatics.