暂无分享,去创建一个
Goran Glavas | Anna Korhonen | Edoardo Maria Ponti | Robert Litschko | Ivan Vulic | A. Korhonen | Goran Glavas | Ivan Vulic | E. Ponti | Robert Litschko
[1] Richard Socher,et al. BERT is Not an Interlingua and the Bias of Tokenization , 2019, EMNLP.
[2] Jörg Tiedemann,et al. News from OPUS — A collection of multilingual parallel corpora with tools and interfaces , 2009 .
[3] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[4] Kees van Deemter,et al. What do you mean, BERT? Assessing BERT as a Distributional Semantics Model , 2019, ArXiv.
[5] Goran Glavas,et al. Unsupervised Cross-Lingual Information Retrieval Using Monolingual Data Only , 2018, SIGIR.
[6] Graham Neubig,et al. XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization , 2020, ICML.
[7] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[8] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[9] Kawin Ethayarajh,et al. How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT, ELMo, and GPT-2 Embeddings , 2019, EMNLP.
[10] Ruize Wang,et al. K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters , 2020, ArXiv.
[11] Daniel Edmiston,et al. A Systematic Analysis of Morphological Content in BERT Models for Multiple Languages , 2020, ArXiv.
[12] Yun-Nung Chen,et al. What Does This Word Mean? Explaining Contextualized Embeddings with Natural Language Definition , 2019, EMNLP.
[13] Marcin Junczys-Dowmunt,et al. The United Nations Parallel Corpus v1.0 , 2016, LREC.
[14] Samuel L. Smith,et al. Offline bilingual word vectors, orthogonal transformations and the inverted softmax , 2017, ICLR.
[15] Geoffrey E. Hinton,et al. Similarity of Neural Network Representations Revisited , 2019, ICML.
[16] Matthew Henderson,et al. Efficient Intent Detection with Dual Sentence Encoders , 2020, NLP4CONVAI.
[17] Qianchu Liu,et al. XCOPA: A Multilingual Dataset for Causal Commonsense Reasoning , 2020, EMNLP.
[18] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[19] Zhiyuan Liu,et al. Understanding the Behaviors of BERT in Ranking , 2019, ArXiv.
[20] Yoshua Bengio,et al. BilBOWA: Fast Bilingual Distributed Representations without Word Alignments , 2014, ICML.
[21] Hinrich Schütze,et al. Generating Derivational Morphology with BERT , 2020, ArXiv.
[22] Wenpeng Yin,et al. Learning Word Meta-Embeddings , 2016, ACL.
[23] Philipp Koehn,et al. Europarl: A Parallel Corpus for Statistical Machine Translation , 2005, MTSUMMIT.
[24] Anders Søgaard,et al. On the Limitations of Unsupervised Bilingual Dictionary Induction , 2018, ACL.
[25] Xuanjing Huang,et al. Pre-trained Models for Natural Language Processing: A Survey , 2020, ArXiv.
[26] Qianchu Liu,et al. Investigating Cross-Lingual Alignment Methods for Contextualized Embeddings with Token-Level Evaluation , 2019, CoNLL.
[27] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[28] Ryan Cotterell,et al. Information-Theoretic Probing for Linguistic Structure , 2020, ACL.
[29] Felix Hill,et al. SimLex-999: Evaluating Semantic Models With (Genuine) Similarity Estimation , 2014, CL.
[30] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[31] Sebastian Ruder,et al. MultiFiT: Efficient Multi-lingual Language Model Fine-tuning , 2019, EMNLP/IJCNLP.
[32] Martin Wattenberg,et al. Visualizing and Measuring the Geometry of BERT , 2019, NeurIPS.
[33] Christopher D. Manning,et al. Finding Universal Grammatical Relations in Multilingual BERT , 2020, ACL.
[34] Anna Rumshisky,et al. A Primer in BERTology: What We Know About How BERT Works , 2020, Transactions of the Association for Computational Linguistics.
[35] Anna Korhonen,et al. On the Relation between Linguistic Typology and (Limitations of) Multilingual Language Modeling , 2018, EMNLP.
[36] Dipanjan Das,et al. BERT Rediscovers the Classical NLP Pipeline , 2019, ACL.
[37] Goran Glavas,et al. Discriminating between Lexico-Semantic Relations with the Specialization Tensor Model , 2018, NAACL.
[38] Danushka Bollegala,et al. Learning Word Meta-Embeddings by Autoencoding , 2018, COLING.
[39] Ivan Titov,et al. Information-Theoretic Probing with Minimum Description Length , 2020, EMNLP.
[40] Eneko Agirre,et al. A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings , 2018, ACL.
[41] Mark Dredze,et al. Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT , 2019, EMNLP.
[42] Yijia Liu,et al. Cross-Lingual BERT Transformation for Zero-Shot Dependency Parsing , 2019, EMNLP.
[43] Goran Glavas,et al. Evaluating Resource-Lean Cross-Lingual Embedding Models in Unsupervised Retrieval , 2019, SIGIR.
[44] Sebastian Ruder,et al. A survey of cross-lingual embedding models , 2017, ArXiv.
[45] Quoc V. Le,et al. Exploiting Similarities among Languages for Machine Translation , 2013, ArXiv.
[46] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[47] Satoshi Matsuoka,et al. Word Embeddings, Analogies, and Machine Learning: Beyond king - man + woman = queen , 2016, COLING.
[48] Christopher D. Manning,et al. A Structural Probe for Finding Syntax in Word Representations , 2019, NAACL.
[49] Joakim Nivre,et al. Do Neural Language Models Show Preferences for Syntactic Formalisms? , 2020, ACL.
[50] Yonatan Belinkov,et al. Linguistic Knowledge and Transferability of Contextual Representations , 2019, NAACL.
[51] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[52] Veselin Stoyanov,et al. Emerging Cross-lingual Structure in Pretrained Language Models , 2020, ACL.
[53] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[54] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[55] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[56] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[57] Goran Glavas,et al. How to (Properly) Evaluate Cross-Lingual Word Embeddings: On Strong Baselines, Comparative Analyses, and Some Misconceptions , 2019, ACL.
[58] Iryna Gurevych,et al. MAD-X: An Adapter-based Framework for Multi-task Cross-lingual Transfer , 2020, EMNLP.
[59] Kyunghyun Cho,et al. Dynamic Meta-Embeddings for Improved Sentence Representations , 2018, EMNLP.
[60] Naveen Arivazhagan,et al. Small and Practical BERT Models for Sequence Labeling , 2019, EMNLP.
[61] Mikel Artetxe,et al. On the Cross-lingual Transferability of Monolingual Representations , 2019, ACL.
[62] Benoît Sagot,et al. What Does BERT Learn about the Structure of Language? , 2019, ACL.
[63] Yonatan Belinkov,et al. Analysis Methods in Neural Language Processing: A Survey , 2018, TACL.