暂无分享,去创建一个
[1] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[2] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[3] Jianfeng Gao,et al. A Human Generated MAchine Reading COmprehension Dataset , 2018 .
[4] W. Bruce Croft,et al. A Deep Relevance Matching Model for Ad-hoc Retrieval , 2016, CIKM.
[5] Chandra Bhagavatula,et al. Semi-supervised sequence tagging with bidirectional language models , 2017, ACL.
[6] Andrew Yates,et al. Contextualized PACRR for Complex Answer Retrieval , 2017, TREC.
[7] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[8] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[9] Kyunghyun Cho,et al. SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine , 2017, ArXiv.
[10] William W. Cohen,et al. Quasar: Datasets for Question Answering by Search and Reading , 2017, ArXiv.
[11] Nick Craswell,et al. Learning to Match using Local and Distributed Representations of Text for Web Search , 2016, WWW.
[12] Zhiyuan Liu,et al. End-to-End Neural Ad-hoc Ranking with Kernel Pooling , 2017, SIGIR.
[13] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[14] Jimmy J. Lin,et al. Anserini: Reproducible Ranking Baselines Using Lucene , 2018, ACM J. Data Inf. Qual..
[15] Gerard de Melo,et al. Co-PACRR: A Context-Aware Neural IR Model for Ad-hoc Retrieval , 2017, WSDM.
[16] Christopher Clark,et al. Simple and Effective Multi-Paragraph Reading Comprehension , 2017, ACL.
[17] Filip Radlinski,et al. TREC Complex Answer Retrieval Overview , 2018, TREC.
[18] Quoc V. Le,et al. QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension , 2018, ICLR.
[19] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[20] Zhiyuan Liu,et al. Convolutional Neural Networks for Soft-Matching N-Grams in Ad-hoc Search , 2018, WSDM.
[21] Jimmy J. Lin,et al. The Neural Hype and Comparisons Against Weak Baselines , 2019, SIGIR Forum.
[22] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.