暂无分享,去创建一个
Yulia Tsvetkov | Zachary C. Lipton | Zirui Wang | Zachary Chase Lipton | Yulia Tsvetkov | Zirui Wang
[1] Qiang Yang,et al. An Overview of Multi-task Learning , 2018 .
[2] Ankur Bapna,et al. Simple, Scalable Adaptation for Neural Machine Translation , 2019, EMNLP.
[3] Jungo Kasai,et al. Polyglot Contextual Representations Improve Crosslingual Transfer , 2019, NAACL.
[4] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[5] Thomas G. Dietterich,et al. To transfer or not to transfer , 2005, NIPS 2005.
[6] Razvan Pascanu,et al. Meta-Learning with Warped Gradient Descent , 2020, ICLR.
[7] Yuji Matsumoto,et al. Universal Dependencies 2.1 , 2017 .
[8] Eva Schlinger,et al. How Multilingual is Multilingual BERT? , 2019, ACL.
[9] Mark Dredze,et al. Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT , 2019, EMNLP.
[10] Jing Gao,et al. On handling negative transfer and imbalanced distributions in multiple source transfer learning , 2014, SDM.
[11] Jaime G. Carbonell,et al. Towards more Reliable Transfer Learning , 2018, ECML/PKDD.
[12] Patrice Marcotte,et al. An overview of bilevel optimization , 2007, Ann. Oper. Res..
[13] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[14] Mark W. Schmidt,et al. Online Learning Rate Adaptation with Hypergradient Descent , 2017, ICLR.
[15] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[16] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[17] Ankur Bapna,et al. Massively Multilingual Neural Machine Translation in the Wild: Findings and Challenges , 2019, ArXiv.
[18] Mikel Artetxe,et al. On the Cross-lingual Transferability of Monolingual Representations , 2019, ACL.
[19] Andrea Vedaldi,et al. Learning multiple visual domains with residual adapters , 2017, NIPS.
[20] Dan Roth,et al. Cross-Lingual Ability of Multilingual BERT: An Empirical Study , 2019, ICLR.
[21] Qiang Yang,et al. A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[22] Karol Hausman,et al. Gradient Surgery for Multi-Task Learning , 2020, NeurIPS.
[23] Geoffrey E. Hinton,et al. Similarity of Neural Network Representations Revisited , 2019, ICML.
[24] Heng Ji,et al. Cross-lingual Name Tagging and Linking for 282 Languages , 2017, ACL.
[25] Mona Attariyan,et al. Parameter-Efficient Transfer Learning for NLP , 2019, ICML.
[26] Martin Wattenberg,et al. Google’s Multilingual Neural Machine Translation System: Enabling Zero-Shot Translation , 2016, TACL.
[27] Ming Zhou,et al. Unicoder: A Universal Language Encoder by Pre-training with Multiple Cross-lingual Tasks , 2019, EMNLP.
[28] Anders Søgaard,et al. On the Limitations of Unsupervised Bilingual Dictionary Induction , 2018, ACL.
[29] Ankur Bapna,et al. Investigating Multilingual NMT Representations at Scale , 2019, EMNLP.
[30] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[31] Jascha Sohl-Dickstein,et al. SVCCA: Singular Vector Canonical Correlation Analysis for Deep Learning Dynamics and Interpretability , 2017, NIPS.
[32] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[33] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[34] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[35] Yulia Tsvetkov,et al. Balancing Training for Multilingual Neural Machine Translation , 2020, ACL.
[36] Graham Neubig,et al. XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization , 2020, ICML.
[37] Jaime G. Carbonell,et al. Characterizing and Avoiding Negative Transfer , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[38] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[39] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[40] Eunsol Choi,et al. TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages , 2020, Transactions of the Association for Computational Linguistics.
[41] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[42] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[43] Veselin Stoyanov,et al. Emerging Cross-lingual Structure in Pretrained Language Models , 2020, ACL.
[44] Dan Klein,et al. Multilingual Alignment of Contextual Word Representations , 2020, ICLR.
[45] Di He,et al. Multilingual Neural Machine Translation with Knowledge Distillation , 2019, ICLR.
[46] Sebastian Ruder,et al. An Overview of Multi-Task Learning in Deep Neural Networks , 2017, ArXiv.
[47] Orhan Firat,et al. Massively Multilingual Neural Machine Translation , 2019, NAACL.
[48] Yiming Yang,et al. Cross-lingual Alignment vs Joint Training: A Comparative Study and A Simple Unified Framework , 2020, ICLR.