暂无分享,去创建一个
Sebastian Schuster | Sonal Gupta | Mike Lewis | Rushin Shah | M. Lewis | Sebastian Schuster | Rushin Shah | S. Gupta
[1] Mauro Cettolo,et al. WIT3: Web Inventory of Transcribed and Translated Talks , 2012, EAMT.
[2] Guillaume Lample,et al. Cross-lingual Language Model Pretraining , 2019, NeurIPS.
[3] Heng Ji,et al. A Multi-lingual Multi-task Architecture for Low-resource Sequence Labeling , 2018, ACL.
[4] Isabelle Augenstein,et al. Parameter sharing between dependency parsers for related languages , 2018, EMNLP.
[5] Gökhan Tür,et al. (Almost) Zero-Shot Cross-Lingual Spoken Language Understanding , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[6] Shrikanth S. Narayanan,et al. A Multi-task Approach to Learning Multilingual Representations , 2018, ACL.
[7] Haoran Li,et al. Multilingual Seq2seq Training with Similarity Loss for Cross-Lingual Document Classification , 2018, Rep4NLP@ACL.
[8] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[9] Joakim Nivre,et al. 82 Treebanks, 34 Models: Universal Dependency Parsing with Multi-Treebank Models , 2018, CoNLL.
[10] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[11] Philip Resnik,et al. Cross-Language Parser Adaptation between Related Languages , 2008, IJCNLP.
[12] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[13] Orhan Firat,et al. Zero-Shot Cross-lingual Classification Using Multilingual Neural Machine Translation , 2018, ArXiv.
[14] Jaime G. Carbonell,et al. Neural Cross-Lingual Named Entity Recognition with Minimal Resources , 2018, EMNLP.
[15] Yann Dauphin,et al. A Convolutional Encoder Model for Neural Machine Translation , 2016, ACL.
[16] Ruslan Salakhutdinov,et al. Transfer Learning for Sequence Tagging with Hierarchical Recurrent Networks , 2016, ICLR.
[17] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[18] Regina Barzilay,et al. Cross-Lingual Alignment of Contextual Word Embeddings, with Applications to Zero-shot Dependency Parsing , 2019, NAACL.
[19] Ahmed Aly,et al. PyText: A Seamless Path from NLP research to production , 2018, ArXiv.
[20] Anna Korhonen,et al. Semantic Specialization of Distributional Word Vector Spaces using Monolingual and Cross-Lingual Constraints , 2017, TACL.
[21] Murhaf Fares,et al. Word vectors, reuse, and replicability: Towards a community repository of large-text resources , 2017, NODALIDA.
[22] Trevor Cohn,et al. Low Resource Dependency Parsing: Cross-lingual Parameter Sharing in a Neural Network Parser , 2015, ACL.
[23] Stephen D. Mayhew,et al. On the Strength of Character Language Models for Multilingual Named Entity Recognition , 2018, EMNLP.
[24] David Yarowsky,et al. Inducing Multilingual Text Analysis Tools via Robust Projection across Aligned Corpora , 2001, HLT.
[25] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[26] Ngoc Thang Vu. Sequential Convolutional Neural Networks for Slot Filling in Spoken Language Understanding , 2016, INTERSPEECH.
[27] Jörg Tiedemann,et al. OpenSubtitles2018: Statistical Rescoring of Sentence Alignments in Large, Noisy Parallel Corpora , 2018, LREC.
[28] Jungo Kasai,et al. Polyglot Contextual Representations Improve Crosslingual Transfer , 2019, NAACL.
[29] David Grangier,et al. Vocabulary Selection Strategies for Neural Machine Translation , 2016, ArXiv.
[30] Barbara Plank,et al. Distant Supervision from Disparate Sources for Low-Resource Part-of-Speech Tagging , 2018, EMNLP.
[31] Richard Socher,et al. Learned in Translation: Contextualized Word Vectors , 2017, NIPS.
[32] Yoshua Bengio,et al. Investigation of recurrent-neural-network architectures and learning methods for spoken language understanding , 2013, INTERSPEECH.
[33] Xin Wang,et al. XL-NBT: A Cross-lingual Neural Belief Tracking Framework , 2018, EMNLP.
[34] Philipp Koehn,et al. Europarl: A Parallel Corpus for Statistical Machine Translation , 2005, MTSUMMIT.
[35] Noah A. Smith,et al. Many Languages, One Parser , 2016, TACL.
[36] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[37] Yijia Liu,et al. Towards Better UD Parsing: Deep Contextualized Word Embeddings, Ensemble, and Treebank Concatenation , 2018, CoNLL.
[38] Wei Lu,et al. Neural Architectures for Multilingual Semantic Parsing , 2017, ACL.
[39] Guillaume Lample,et al. Neural Architectures for Named Entity Recognition , 2016, NAACL.
[40] Timothy Dozat,et al. Stanford’s Graph-based Neural Dependency Parser at the CoNLL 2017 Shared Task , 2017, CoNLL.
[41] Guillaume Lample,et al. Word Translation Without Parallel Data , 2017, ICLR.
[42] Anders Søgaard,et al. A Survey of Cross-lingual Word Embedding Models , 2017, J. Artif. Intell. Res..
[43] Bing Liu,et al. Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling , 2016, INTERSPEECH.
[44] Slav Petrov,et al. Multi-Source Transfer of Delexicalized Dependency Parsers , 2011, EMNLP.
[45] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.