How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models
暂无分享,去创建一个
[1] Frank Hutter,et al. Decoupled Weight Decay Regularization , 2017, ICLR.
[2] Dirk Hovy,et al. What the [MASK]? Making Sense of Language-Specific BERT Models , 2020, ArXiv.
[3] Andrea Vedaldi,et al. Learning multiple visual domains with residual adapters , 2017, NIPS.
[4] Sampo Pyysalo,et al. Universal Dependencies v2: An Evergrowing Multilingual Treebank Collection , 2020, LREC.
[5] Jan Hajic,et al. UDPipe: Trainable Pipeline for Processing CoNLL-U Files Performing Tokenization, Morphological Analysis, POS Tagging and Parsing , 2016, LREC.
[6] Iryna Gurevych,et al. Common Sense or World Knowledge? Investigating Adapter-Based Knowledge Injection into Pretrained Transformers , 2020, DEELIO.
[7] Gertjan van Noord,et al. UDapter: Language Adaptation for Truly Universal Dependency Parsing , 2020, EMNLP.
[8] Hazem Hajj,et al. AraBERT: Transformer-based Model for Arabic Language Understanding , 2020, OSACT.
[9] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[10] Miikka Silfverberg,et al. A Finnish news corpus for named entity recognition , 2019, Language Resources and Evaluation.
[11] Sergey Smetanin,et al. Sentiment Analysis of Product Reviews in Russian using Convolutional Neural Networks , 2019, 2019 IEEE 21st Conference on Business Informatics (CBI).
[12] Colin Raffel,et al. mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer , 2021, NAACL.
[13] Ayu Purwarianti,et al. Improving Bi-LSTM Performance for Indonesian Sentiment Analysis Using Paragraph Vector , 2019, 2019 International Conference of Advanced Informatics: Concepts, Theory and Applications (ICAICTA).
[14] Goran Glavas,et al. Is Supervised Syntactic Parsing Beneficial for Language Understanding Tasks? An Empirical Investigation , 2020, EACL.
[15] Mark Dredze,et al. Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT , 2019, EMNLP.
[16] Heng Ji,et al. Cross-lingual Name Tagging and Linking for 282 Languages , 2017, ACL.
[17] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[18] Thomas Eckart,et al. OSIAN: Open Source International Arabic News Corpus - Preparation and Integration into the CLARIN-infrastructure , 2019, WANLP@ACL 2019.
[19] John Wieting,et al. CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation , 2021, ArXiv.
[20] Goran Glavaš,et al. From Zero to Hero: On the Limitations of Zero-Shot Language Transfer with Multilingual Transformers , 2020, EMNLP.
[21] Iryna Gurevych,et al. AdapterFusion: Non-Destructive Task Composition for Transfer Learning , 2021, EACL.
[22] Iain Murray,et al. BERT and PALs: Projected Attention Layers for Efficient Adaptation in Multi-Task Learning , 2019, ICML.
[23] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[24] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[25] Graham Neubig,et al. XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization , 2020, ICML.
[26] Leonid Boytsov,et al. SberQuAD - Russian Reading Comprehension Dataset: Description and Analysis , 2020, CLEF.
[27] Samuel R. Bowman,et al. When Do You Need Billions of Words of Pretraining Data? , 2020, ACL.
[28] Giovanni Semeraro,et al. AlBERTo: Italian BERT Language Understanding Model for NLP Challenging Tasks Based on Tweets , 2019, CLiC-it.
[29] Sanja Fidler,et al. Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[30] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[31] Ayu Purwarianti,et al. IndoNLU: Benchmark and Resources for Evaluating Indonesian Natural Language Understanding , 2020, AACL.
[32] Mikel Artetxe,et al. On the Cross-lingual Transferability of Monolingual Representations , 2019, ACL.
[33] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[34] Mark Dredze,et al. Are All Languages Created Equal in Multilingual BERT? , 2020, REPL4NLP.
[35] Sampo Pyysalo,et al. WikiBERT Models: Deep Transfer Learning for Many Languages , 2020, NODALIDA.
[36] Mikhail Arkhipov,et al. Adaptation of Deep Bidirectional Multilingual Transformers for Russian Language , 2019, ArXiv.
[37] Trevor Cohn,et al. Massively Multilingual Transfer for NER , 2019, ACL.
[38] Xu Sun,et al. A Discourse-Level Named Entity Recognition and Relation Extraction Dataset for Chinese Literature Text , 2017, ArXiv.
[39] Nadir Durrani,et al. Farasa: A Fast and Furious Segmenter for Arabic , 2016, NAACL.
[40] Iryna Gurevych,et al. MultiCQA: Zero-Shot Transfer of Self-Supervised Text Matching Models on a Massive Scale , 2020, EMNLP.
[41] Jungo Kasai,et al. Polyglot Contextual Representations Improve Crosslingual Transfer , 2019, NAACL.
[42] Mykola Pechenizkiy,et al. Cross-lingual polarity detection with machine translation , 2013, WISDOM '13.
[43] Enkhbold Bataa,et al. An Investigation of Transfer Learning-Based Sentiment Analysis in Japanese , 2019, ACL.
[44] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[45] Anna Korhonen,et al. On the Relation between Linguistic Typology and (Limitations of) Multilingual Language Modeling , 2018, EMNLP.
[46] Iryna Gurevych,et al. UNKs Everywhere: Adapting Multilingual Language Models to New Scripts , 2021, EMNLP.
[47] Iryna Gurevych,et al. MAD-X: An Adapter-based Framework for Multi-task Cross-lingual Transfer , 2020, EMNLP.
[48] Noah A. Smith,et al. Parsing with Multilingual BERT, a Small Treebank, and a Small Corpus , 2020, FINDINGS.
[49] Goran Glavas,et al. Probing Pretrained Language Models for Lexical Semantics , 2020, EMNLP.
[50] A. Elnagar,et al. Hotel Arabic-Reviews Dataset Construction for Sentiment Analysis Applications , 2018 .
[51] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[52] Francisco Casacuberta,et al. How Much Does Tokenization Affect Neural Machine Translation? , 2018, CICLing.
[53] Seungyoung Lim,et al. KorQuAD1.0: Korean QA Dataset for Machine Reading Comprehension , 2019, ArXiv.
[54] Laurent Romary,et al. A Monolingual Approach to Contextualized Word Embeddings for Mid-Resource Languages , 2020, ACL.
[55] Stefan Schweter,et al. BERTurk - BERT models for Turkish , 2020 .
[56] Lysandre Debut,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[57] Sangah Lee,et al. KR-BERT: A Small-Scale Korean-Specific Language Model , 2020, 2008.03979.
[58] Sebastian Riedel,et al. MLQA: Evaluating Cross-lingual Extractive Question Answering , 2019, ACL.
[59] Monojit Choudhury,et al. The State and Fate of Linguistic Diversity and Inclusion in the NLP World , 2020, ACL.
[60] Yuting Lai,et al. DRCD: a Chinese Machine Reading Comprehension Dataset , 2018, ArXiv.
[61] Laurent Romary,et al. CamemBERT: a Tasty French Language Model , 2019, ACL.
[62] Noah A. Smith,et al. Parsing with Multilingual BERT, a Small Corpus, and a Small Treebank , 2020, EMNLP 2020.
[63] Hyung Won Chung,et al. Improving Multilingual Models with Language-Clustered Vocabularies , 2020, EMNLP.
[64] Qianchu Liu,et al. XCOPA: A Multilingual Dataset for Causal Commonsense Reasoning , 2020, EMNLP.
[65] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[66] Iryna Gurevych,et al. AdapterHub: A Framework for Adapting Transformers , 2020, EMNLP.
[67] Tapio Salakoski,et al. Multilingual is not enough: BERT for Finnish , 2019, ArXiv.
[68] Tommaso Caselli,et al. BERTje: A Dutch BERT Model , 2019, ArXiv.
[69] Dan Roth,et al. Cross-Lingual Ability of Multilingual BERT: An Empirical Study , 2019, ICLR.
[70] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[71] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[72] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[73] Iryna Gurevych,et al. AdapterDrop: On the Efficiency of Adapters in Transformers , 2020, EMNLP.
[74] Eunsol Choi,et al. TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages , 2020, Transactions of the Association for Computational Linguistics.
[75] Mona Attariyan,et al. Parameter-Efficient Transfer Learning for NLP , 2019, ICML.
[76] Thierry Poibeau,et al. Modeling Language Variation and Universals: A Survey on Typological Linguistics for Natural Language Processing , 2018, Computational Linguistics.
[77] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[78] Tapio Salakoski,et al. Is Multilingual BERT Fluent in Language Generation? , 2019, ArXiv.