暂无分享,去创建一个
Benjamin Lecouteux | Alexandre Allauzen | Laurent Besacier | Maximin Coavoux | Didier Schwab | Vincent Segonne | Jibril Frej | Benoit Crabb'e | Hang Le | Loic Vial
[1] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[2] Jörg Tiedemann,et al. OpenSubtitles2016: Extracting Large Parallel Corpora from Movie and TV Subtitles , 2016, LREC.
[3] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[4] Dan Klein,et al. Constituency Parsing with a Self-Attentive Encoder , 2018, ACL.
[5] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[6] Pieter Delobelle,et al. RobBERT: a Dutch RoBERTa-based Language Model , 2020, EMNLP.
[7] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[8] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[9] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[10] Kilian Q. Weinberger,et al. Deep Networks with Stochastic Depth , 2016, ECCV.
[11] John Blitzer,et al. Domain Adaptation with Structural Correspondence Learning , 2006, EMNLP.
[12] Tommaso Caselli,et al. BERTje: A Dutch BERT Model , 2019, ArXiv.
[13] George A. Miller,et al. WordNet: A Lexical Database for English , 1995, HLT.
[14] Jason Baldridge,et al. PAWS-X: A Cross-lingual Adversarial Dataset for Paraphrase Identification , 2019, EMNLP.
[15] Julian Salazar,et al. Transformers without Tears: Improving the Normalization of Self-Attention , 2019, ArXiv.
[16] Marie Candito,et al. Using Wiktionary as a resource for WSD : the case of French verbs , 2019, IWCS.
[17] Dan Klein,et al. Multilingual Constituency Parsing with Self-Attention and Pre-Training , 2018, ACL.
[18] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[19] Marie Candito,et al. The LIGM-Alpage architecture for the SPMRL 2013 Shared Task: Multiword Expression Analysis and Dependency Parsing , 2013, SPMRL@EMNLP.
[20] George A. Miller,et al. A Semantic Concordance , 1993, HLT.
[21] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[22] Tapio Salakoski,et al. Multilingual is not enough: BERT for Finnish , 2019, ArXiv.
[23] Philipp Koehn,et al. Europarl: A Parallel Corpus for Statistical Machine Translation , 2005, MTSUMMIT.
[24] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[25] Alexandra Kinyon,et al. Building a Treebank for French , 2000, LREC.
[26] Nizar Habash,et al. Overview of the SPMRL 2013 Shared Task: A Cross-Framework Evaluation of Parsing Morphologically Rich Languages , 2013, SPMRL@EMNLP.
[27] Jan Niehues,et al. Very Deep Self-Attention Networks for End-to-End Speech Recognition , 2019, INTERSPEECH.
[28] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[29] Tengyu Ma,et al. Fixup Initialization: Residual Learning Without Normalization , 2019, ICLR.
[30] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[31] Benjamin Lecouteux,et al. UFSAC: Unification of Sense Annotated Corpora and Tools , 2018, LREC.
[32] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[33] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[34] Rodrigo Nogueira,et al. Portuguese Named Entity Recognition using BERT-CRF , 2019, ArXiv.
[35] Benno Stein,et al. Cross-Language Text Classification Using Structural Correspondence Learning , 2010, ACL.
[36] Mikhail Arkhipov,et al. Adaptation of Deep Bidirectional Multilingual Transformers for Russian Language , 2019, ArXiv.
[37] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[38] Jörg Tiedemann,et al. Billions of Parallel Words for Free: Building and Using the EU Bookshop Corpus , 2014, LREC.
[39] R'emi Louf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[40] Hazem M. Hajj,et al. AraBERT: Transformer-based Model for Arabic Language Understanding , 2020, OSACT.
[41] Gosse Bouma,et al. 48th Annual Meeting of the Association for Computational Linguistics , 2010, ACL 2010.
[42] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[43] Yonatan Belinkov,et al. Findings of the First Shared Task on Machine Translation Robustness , 2019, WMT.
[44] Benjamin Lecouteux,et al. Sense Vocabulary Compression through the Semantic Knowledge of WordNet for Neural Word Sense Disambiguation , 2019, GWC.
[45] Giovanni Semeraro,et al. AlBERTo: Italian BERT Language Understanding Model for NLP Challenging Tasks Based on Tweets , 2019, CLiC-it.
[46] Andreas Eisele,et al. MultiUN: A Multilingual Corpus from United Nation Documents , 2010, LREC.
[47] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[48] Di He,et al. Efficient Training of BERT by Progressively Stacking , 2019, ICML.
[49] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[50] Simone Paolo Ponzetto,et al. BabelNet: Building a Very Large Multilingual Semantic Network , 2010, ACL.
[51] Jingbo Zhu,et al. Learning Deep Transformer Models for Machine Translation , 2019, ACL.
[52] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[53] Zhen-Hua Ling,et al. Enhanced LSTM for Natural Language Inference , 2016, ACL.
[54] Jason Weston,et al. A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.
[55] Laurent Romary,et al. CamemBERT: a Tasty French Language Model , 2019, ACL.
[56] Gilles Sérasset,et al. Dbnary: Wiktionary as a LMF based Multilingual RDF network , 2012, LREC.
[57] Colin Raffel,et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..
[58] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[59] Edouard Grave,et al. Reducing Transformer Depth on Demand with Structured Dropout , 2019, ICLR.
[60] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[61] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[62] Kenneth Heafield,et al. N-gram Counts and Language Models from the Common Crawl , 2014, LREC.
[63] Quoc V. Le,et al. Unsupervised Pretraining for Sequence to Sequence Learning , 2016, EMNLP.
[64] Holger Schwenk,et al. Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond , 2018, Transactions of the Association for Computational Linguistics.
[65] Roberto Navigli,et al. Word sense disambiguation: A survey , 2009, CSUR.
[66] Roberto Navigli,et al. SemEval-2013 Task 12: Multilingual Word Sense Disambiguation , 2013, *SEMEVAL.
[67] Jörg Tiedemann,et al. Parallel Data, Tools and Interfaces in OPUS , 2012, LREC.
[68] Samy Bengio,et al. Tensor2Tensor for Neural Machine Translation , 2018, AMTA.
[69] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[70] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[71] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[72] Omer Levy,et al. SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems , 2019, NeurIPS.
[73] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[74] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[75] Sebastian Ruder,et al. MultiFiT: Efficient Multi-lingual Language Model Fine-tuning , 2019, EMNLP/IJCNLP.
[76] Dat Quoc Nguyen,et al. PhoBERT: Pre-trained language models for Vietnamese , 2020, EMNLP.
[77] Jason Baldridge,et al. PAWS: Paraphrase Adversaries from Word Scrambling , 2019, NAACL.