Parsing with Multilingual BERT, a Small Treebank, and a Small Corpus
暂无分享,去创建一个
[1] Eneko Agirre,et al. Give your Text Representation Models some Love: the Case for Basque , 2020, LREC.
[2] Yue Zhang,et al. Universal Dependencies Parsing for Colloquial Singaporean English , 2017, ACL.
[3] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[4] Ming-Wei Chang,et al. Zero-Shot Entity Linking by Reading Entity Descriptions , 2019, ACL.
[5] Jungo Kasai,et al. Polyglot Contextual Representations Improve Crosslingual Transfer , 2019, NAACL.
[6] Deniz Yuret,et al. Transfer Learning for Low-Resource Neural Machine Translation , 2016, EMNLP.
[7] Tapio Salakoski,et al. Is Multilingual BERT Fluent in Language Generation? , 2019, ArXiv.
[8] Daniel Kondratyuk,et al. 75 Languages, 1 Model: Parsing Universal Dependencies Universally , 2019, EMNLP.
[9] Luke S. Zettlemoyer,et al. AllenNLP: A Deep Semantic Natural Language Processing Platform , 2018, ArXiv.
[10] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[11] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[12] Xuanjing Huang,et al. How to Fine-Tune BERT for Text Classification? , 2019, CCL.
[13] Eva Schlinger,et al. How Multilingual is Multilingual BERT? , 2019, ACL.
[14] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[15] Mark Dredze,et al. Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT , 2019, EMNLP.
[16] Regina Barzilay,et al. Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-shot Dependency Parsing , 2019, NAACL.
[17] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[18] Peter Trudgill,et al. A Glossary of Sociolinguistics , 2019 .
[19] Wei-Hung Weng,et al. Publicly Available Clinical BERT Embeddings , 2019, Proceedings of the 2nd Clinical Natural Language Processing Workshop.
[20] Kathleen McKeown,et al. IMHO Fine-Tuning Improves Claim Detection , 2019, NAACL.
[21] Veselin Stoyanov,et al. Emerging Cross-lingual Structure in Pretrained Language Models , 2020, ACL.
[22] Dan Klein,et al. Multilingual Alignment of Contextual Word Representations , 2020, ICLR.
[23] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[24] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[25] Iz Beltagy,et al. SciBERT: A Pretrained Language Model for Scientific Text , 2019, EMNLP.
[26] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[27] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[28] Jacob Eisenstein,et al. Unsupervised Domain Adaptation of Contextualized Embeddings for Sequence Labeling , 2019, EMNLP.
[29] Dan Roth,et al. Cross-Lingual Ability of Multilingual BERT: An Empirical Study , 2019, ICLR.
[30] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[31] Jungo Kasai,et al. Low-Resource Parsing with Crosslingual Contextualized Representations , 2019, CoNLL.
[32] Marcello Federico,et al. Transfer Learning in Multilingual Neural Machine Translation with Dynamic Vocabulary , 2018, IWSLT.
[33] Mark Dredze,et al. Sources of Transfer in Multilingual Named Entity Recognition , 2020, ACL.
[34] Mark Dredze,et al. Are All Languages Created Equal in Multilingual BERT? , 2020, REPL4NLP.
[35] Naveen Arivazhagan,et al. Small and Practical BERT Models for Sequence Labeling , 2019, EMNLP.
[36] Mikel Artetxe,et al. On the Cross-lingual Transferability of Monolingual Representations , 2019, ACL.
[37] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[38] Roy Schwartz,et al. Show Your Work: Improved Reporting of Experimental Results , 2019, EMNLP.