暂无分享,去创建一个
[1] Wei Zhang,et al. Evidence Aggregation for Answer Re-Ranking in Open-Domain Question Answering , 2017, ICLR.
[2] Sergei Vassilvitskii,et al. k-means++: the advantages of careful seeding , 2007, SODA '07.
[3] Guillaume Lample,et al. Neural Architectures for Named Entity Recognition , 2016, NAACL.
[4] Ali Farhadi,et al. Phrase-Indexed Question Answering: A New Challenge for Scalable Document Comprehension , 2018, EMNLP.
[5] James H. Martin,et al. Speech and Language Processing, 2nd Edition , 2008 .
[6] Sebastian Riedel,et al. Constructing Datasets for Multi-hop Reading Comprehension Across Documents , 2017, TACL.
[7] Christopher Clark,et al. Simple and Effective Multi-Paragraph Reading Comprehension , 2017, ACL.
[8] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[9] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[10] Wei Zhang,et al. R3: Reinforced Ranker-Reader for Open-Domain Question Answering , 2018, AAAI.
[11] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[12] Richard Socher,et al. Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering , 2019, ICLR.
[13] Luo Si,et al. A Deep Cascade Model for Multi-Document Reading Comprehension , 2018, AAAI.
[14] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[15] Ronen Feldman,et al. Using Corpus Statistics on Entities to Improve Semi-supervised Relation Extraction from the Web , 2007, ACL.
[16] Jonathan Berant,et al. The Web as a Knowledge-Base for Answering Complex Questions , 2018, NAACL.
[17] William W. Cohen,et al. Quasar: Datasets for Question Answering by Search and Reading , 2017, ArXiv.
[18] Stefan Feuerriegel,et al. Adaptive Document Retrieval for Deep Question Answering , 2018, EMNLP.
[19] Samuel R. Bowman,et al. Training a Ranking Function for Open-Domain Question Answering , 2018, NAACL.
[20] Mihai Surdeanu,et al. The Stanford CoreNLP Natural Language Processing Toolkit , 2014, ACL.
[21] Yoshua Bengio,et al. HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering , 2018, EMNLP.
[22] Jason Weston,et al. Key-Value Memory Networks for Directly Reading Documents , 2016, EMNLP.
[23] Zhiyuan Liu,et al. Denoising Distantly Supervised Open-Domain Question Answering , 2018, ACL.
[24] Pascal Vincent,et al. Hierarchical Memory Networks , 2016, ArXiv.
[25] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[26] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[27] Jeff Johnson,et al. Billion-Scale Similarity Search with GPUs , 2017, IEEE Transactions on Big Data.
[28] Jason Weston,et al. Memory Networks , 2014, ICLR.
[29] Ruslan Salakhutdinov,et al. Neural Models for Reasoning over Multiple Mentions Using Coreference , 2018, NAACL.
[30] Ellen M. Voorhees,et al. Building a question answering test collection , 2000, SIGIR '00.
[31] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.
[32] Kyunghyun Cho,et al. SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine , 2017, ArXiv.
[33] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[34] Danqi Chen,et al. Reasoning With Neural Tensor Networks for Knowledge Base Completion , 2013, NIPS.
[35] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[36] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[37] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[38] Rajarshi Das,et al. Multi-step Retriever-Reader Interaction for Scalable Open-domain Question Answering , 2019, ICLR.
[39] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[40] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[41] Jason Weston,et al. End-To-End Memory Networks , 2015, NIPS.
[42] Jianfeng Gao,et al. A Human Generated MAchine Reading COmprehension Dataset , 2018 .
[43] Ronen Feldman,et al. Techniques and applications for sentiment analysis , 2013, CACM.
[44] Jimmy J. Lin,et al. End-to-End Open-Domain Question Answering with BERTserini , 2019, NAACL.
[45] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[46] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[47] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[48] Jaewoo Kang,et al. Ranking Paragraphs for Improving Answer Recall in Open-Domain Question Answering , 2018, EMNLP.
[49] J. Ross Quinlan,et al. Induction of Decision Trees , 1986, Machine Learning.
[50] Richard Socher,et al. Efficient and Robust Question Answering from Minimal Context over Documents , 2018, ACL.
[51] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.