FlauBERT : des modèles de langue contextualisés pré-entraînés pour le français (FlauBERT : Unsupervised Language Model Pre-training for French)
暂无分享,去创建一个
Alexandre Allauzen | Laurent Besacier | Maximin Coavoux | Didier Schwab | Benoît Crabbé | Vincent Segonne | Loïc Vial | Jibril Frej | Benjamin Lecouteux | Hang Le | A. Allauzen | Benoît Crabbé | L. Besacier | D. Schwab | Maximin Coavoux | Loïc Vial | Hang Le | Jibril Frej | V. Segonne | B. Lecouteux
[1] Benno Stein,et al. Cross-Language Text Classification Using Structural Correspondence Learning , 2010, ACL.
[2] George A. Miller,et al. WordNet: A Lexical Database for English , 1995, HLT.
[3] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[4] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[5] Zhen-Hua Ling,et al. Enhanced LSTM for Natural Language Inference , 2016, ACL.
[6] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[7] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[8] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[9] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[10] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[11] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[12] Quoc V. Le,et al. Unsupervised Pretraining for Sequence to Sequence Learning , 2016, EMNLP.
[13] Holger Schwenk,et al. Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond , 2018, Transactions of the Association for Computational Linguistics.
[14] Benjamin Lecouteux,et al. Sense Vocabulary Compression through the Semantic Knowledge of WordNet for Neural Word Sense Disambiguation , 2019, GWC.
[15] Dan Klein,et al. Constituency Parsing with a Self-Attentive Encoder , 2018, ACL.
[16] Benjamin Lecouteux,et al. Compression de vocabulaire de sens grâce aux relations sémantiques pour la désambiguïsation lexicale (Sense Vocabulary Compression through Semantic Knowledge for Word Sense Disambiguation) , 2019, PFIA.
[17] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[18] Marie Candito,et al. Using Wiktionary as a resource for WSD : the case of French verbs , 2019, IWCS.
[19] Sebastian Ruder,et al. MultiFiT: Efficient Multi-lingual Language Model Fine-tuning , 2019, EMNLP/IJCNLP.
[20] Jason Baldridge,et al. PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification , 2019, EMNLP.
[21] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[22] John Blitzer,et al. Domain Adaptation with Structural Correspondence Learning , 2006, EMNLP.
[23] Omer Levy,et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems , 2019, NeurIPS.
[24] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[25] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[26] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[27] Marie Candito,et al. The LIGM-Alpage architecture for the SPMRL 2013 Shared Task: Multiword Expression Analysis and Dependency Parsing , 2013, SPMRL@EMNLP.
[28] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[29] Roberto Navigli,et al. SemEval-2013 Task 12: Multilingual Word Sense Disambiguation , 2013, *SEMEVAL.
[30] Jörg Tiedemann,et al. Parallel Data, Tools and Interfaces in OPUS , 2012, LREC.
[31] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[32] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[33] Simone Paolo Ponzetto,et al. BabelNet: Building a Very Large Multilingual Semantic Network , 2010, ACL.
[34] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[35] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[36] Jason Baldridge,et al. PAWS: Paraphrase Adversaries from Word Scrambling , 2019, NAACL.
[37] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[38] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[39] Nizar Habash,et al. Overview of the SPMRL 2013 Shared Task: A Cross-Framework Evaluation of Parsing Morphologically Rich Languages , 2013, SPMRL@EMNLP.
[40] Dan Klein,et al. Multilingual Constituency Parsing with Self-Attention and Pre-Training , 2018, ACL.