暂无分享,去创建一个
Avirup Sil | Sara Rosenthal | Radu Florian | Mihaela A. Bornea | Lin Pan | Mihaela Bornea | Sara Rosenthal | Radu Florian | Avirup Sil | Lin Pan
[1] Hazem M. Hajj,et al. Neural Arabic Question Answering , 2019, WANLP@ACL 2019.
[2] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[3] Michael Collins,et al. Synthetic QA Corpora Generation with Roundtrip Consistency , 2019, ACL.
[4] Yuting Lai,et al. DRCD: a Chinese Machine Reading Comprehension Dataset , 2018, ArXiv.
[5] Seung-won Hwang,et al. Semi-supervised Training Data Generation for Multilingual Question Answering , 2018, LREC.
[6] Mohit Bansal,et al. Robust Machine Comprehension Models via Adversarial Training , 2018, NAACL.
[7] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[8] Yu Cheng,et al. FreeLB: Enhanced Adversarial Training for Natural Language Understanding , 2020, ICLR.
[9] Jian Li,et al. Multi-Head Attention with Disagreement Regularization , 2018, EMNLP.
[10] Claire Cardie,et al. Adversarial Deep Averaging Networks for Cross-Lingual Sentiment Classification , 2016, TACL.
[11] Roberto Basili,et al. Enabling deep learning for large scale question answering in Italian , 2019, Intelligenza Artificiale.
[12] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[13] Veselin Stoyanov,et al. Unsupervised Cross-lingual Representation Learning at Scale , 2019, ACL.
[14] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[15] Pushpak Bhattacharyya,et al. MMQA: A Multi-domain Multi-lingual Question-Answering Framework for English and Hindi , 2018, LREC.
[16] David Yarowsky,et al. Inducing Multilingual Text Analysis Tools via Robust Projection across Aligned Corpora , 2001, HLT.
[17] Sebastian Riedel,et al. MLQA: Evaluating Cross-lingual Extractive Question Answering , 2019, ACL.
[18] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[19] Yichao Lu,et al. Adversarial Learning with Contextual Embeddings for Zero-resource Cross-lingual Classification and NER , 2019, EMNLP/IJCNLP.
[20] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[21] Wanxiang Che,et al. Improving Machine Reading Comprehension via Adversarial Training , 2019, ArXiv.
[22] Xinyan Xiao,et al. DuReader: a Chinese Machine Reading Comprehension Dataset from Real-world Applications , 2017, QA@ACL.
[23] Daniele Bonadiman,et al. Effective shared representations with Multitask Learning for Community Question Answering , 2017, EACL.
[24] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[25] Yoshua Bengio,et al. HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering , 2018, EMNLP.
[26] Eunsol Choi,et al. TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages , 2020, Transactions of the Association for Computational Linguistics.
[27] Sameer Singh,et al. Universal Adversarial Triggers for Attacking and Analyzing NLP , 2019, EMNLP.
[28] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[29] Richard Socher,et al. The Natural Language Decathlon: Multitask Learning as Question Answering , 2018, ArXiv.
[30] Philip Bachman,et al. NewsQA: A Machine Comprehension Dataset , 2016, Rep4NLP@ACL.
[31] Ankur Bapna,et al. Massively Multilingual Neural Machine Translation in the Wild: Findings and Challenges , 2019, ArXiv.
[32] Wentao Ma,et al. A Span-Extraction Dataset for Chinese Machine Reading Comprehension , 2019, EMNLP-IJCNLP.
[33] Wanxiang Che,et al. Cross-Lingual Machine Reading Comprehension , 2019, EMNLP/IJCNLP.
[34] Mikel Artetxe,et al. On the Cross-lingual Transferability of Monolingual Representations , 2019, ACL.
[35] Guillaume Lample,et al. XNLI: Evaluating Cross-lingual Sentence Representations , 2018, EMNLP.
[36] Quoc V. Le,et al. QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension , 2018, ICLR.
[37] Lucas Theis,et al. Amortised MAP Inference for Image Super-resolution , 2016, ICLR.
[38] Mark Dredze,et al. Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT , 2019, EMNLP.
[39] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[40] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[41] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[42] Nan Duan,et al. Enhancing Answer Boundary Detection for Multilingual Machine Reading Comprehension , 2020, ACL.
[43] Yoshimasa Tsuruoka,et al. Multilingual Extractive Reading Comprehension by Runtime Machine Translation , 2018, ArXiv.