暂无分享,去创建一个
[1] Taku Kudo,et al. SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing , 2018, EMNLP.
[2] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[3] Xinyan Xiao,et al. DuReader: a Chinese Machine Reading Comprehension Dataset from Real-world Applications , 2017, QA@ACL.
[4] Zhen Huang,et al. A Multi-Type Multi-Span Network for Reading Comprehension that Requires Discrete Reasoning , 2019, EMNLP.
[5] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[6] Junji Tomita,et al. A Simple but Effective Method to Incorporate Multi-turn Context with BERT for Conversational Machine Comprehension , 2019, Proceedings of the First Workshop on NLP for Conversational AI.
[7] Christopher Clark,et al. Simple and Effective Multi-Paragraph Reading Comprehension , 2017, ACL.
[8] Mark Yatskar,et al. A Qualitative Comparison of CoQA, SQuAD 2.0 and QuAC , 2018, NAACL.
[9] Jianfeng Gao,et al. A Human Generated MAchine Reading COmprehension Dataset , 2018 .
[10] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[11] Kai Liu,et al. Multi-Passage Machine Reading Comprehension with Cross-Passage Answer Verification , 2018, ACL.
[12] Danqi Chen,et al. CoQA: A Conversational Question Answering Challenge , 2018, TACL.
[13] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[14] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[15] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[16] Seohyun Back,et al. Cut to the Chase: A Context Zoom-in Network for Reading Comprehension , 2018, EMNLP.
[17] Hai Zhao,et al. Dual Multi-head Co-attention for Multi-choice Reading Comprehension , 2020, ArXiv.
[18] Wiebke Wagner,et al. Steven Bird, Ewan Klein and Edward Loper: Natural Language Processing with Python, Analyzing Text with the Natural Language Toolkit , 2010, Lang. Resour. Evaluation.
[19] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[20] Hai Zhao,et al. Machine Reading Comprehension: The Role of Contextualized Language Models and Beyond , 2020, ArXiv.
[21] Hai Zhao,et al. Semantics-aware BERT for Language Understanding , 2020, AAAI.
[22] Thomas Wolf,et al. HuggingFace's Transformers: State-of-the-art Natural Language Processing , 2019, ArXiv.
[23] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[24] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[25] Siu Cheung Hui,et al. Multi-Granular Sequence Encoding via Dilated Compositional Units for Reading Comprehension , 2018, EMNLP.
[26] Hai Zhao,et al. Dual Co-Matching Network for Multi-choice Reading Comprehension , 2020, AAAI.
[27] Chang Zhou,et al. Cognitive Graph for Multi-Hop Reading Comprehension at Scale , 2019, ACL.
[28] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[29] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[30] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[31] Zhuosheng Zhang,et al. SG-Net: Syntax-Guided Machine Reading Comprehension , 2019, AAAI.
[32] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[33] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[34] Luo Si,et al. A Deep Cascade Model for Multi-Document Reading Comprehension , 2018, AAAI.
[35] Mihai Surdeanu,et al. The Stanford CoreNLP Natural Language Processing Toolkit , 2014, ACL.
[36] Lei Li,et al. Dynamically Fused Graph Network for Multi-hop Reasoning , 2019, ACL.
[37] G P Shrivatsa Bhargav,et al. Span Selection Pre-training for Question Answering , 2019, ACL.
[38] Kyunghyun Cho,et al. SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine , 2017, ArXiv.
[39] Ming Zhou,et al. S-Net: From Answer Extraction to Answer Synthesis for Machine Reading Comprehension , 2018, AAAI.
[40] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[41] Hai Zhao,et al. Retrospective Reader for Machine Reading Comprehension , 2020, AAAI.
[42] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[43] Mohit Bansal,et al. Commonsense for Generative Multi-Hop Question Answering Tasks , 2018, EMNLP.
[44] Junji Tomita,et al. Multi-style Generative Reading Comprehension , 2019, ACL.
[45] Wei Zhang,et al. R3: Reinforced Ranker-Reader for Open-Domain Question Answering , 2018, AAAI.
[46] An Yang,et al. Enhancing Pre-Trained Language Representations with Rich Knowledge for Machine Reading Comprehension , 2019, ACL.
[47] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[48] Chris Dyer,et al. The NarrativeQA Reading Comprehension Challenge , 2017, TACL.
[49] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[50] Jian Su,et al. Densely Connected Attention Propagation for Reading Comprehension , 2018, NeurIPS.
[51] Xiang Ren,et al. KagNet: Knowledge-Aware Graph Networks for Commonsense Reasoning , 2019, EMNLP.