暂无分享,去创建一个
[1] Ramesh Nallapati,et al. Multi-passage BERT: A Globally Normalized BERT Model for Open-domain Question Answering , 2019, EMNLP.
[2] Fei Wang,et al. Ieee Transactions on Knowledge and Data Engineering, Manuscropt Id 1 Social Recommendation with Cross-domain Transferable Knowledge , 2022 .
[3] Pradeep Ravikumar,et al. Word Mover’s Embedding: From Word2Vec to Document Embedding , 2018, EMNLP.
[4] Weiming Zhang,et al. Neural Machine Reading Comprehension: Methods and Trends , 2019, Applied Sciences.
[5] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[6] Jonathan Berant,et al. On Making Reading Comprehension More Comprehensive , 2019, EMNLP.
[7] Mohammed J. Zaki,et al. GraphFlow: Exploiting Conversation Flow with Graph Neural Networks for Conversational Machine Comprehension , 2019, IJCAI.
[8] Danqi Chen,et al. Dense Passage Retrieval for Open-Domain Question Answering , 2020, EMNLP.
[9] Wenhan Xiong,et al. Pretrained Encyclopedia: Weakly Supervised Knowledge-Pretrained Language Model , 2019, ICLR.
[10] Wenhao Yu,et al. A Technical Question Answering System with Transfer Learning , 2020, EMNLP.
[11] Ming Zhou,et al. Gated Self-Matching Networks for Reading Comprehension and Question Answering , 2017, ACL.
[12] Mohammed J. Zaki,et al. Bidirectional Attentive Memory Networks for Question Answering over Knowledge Bases , 2019, NAACL.
[13] Avirup Sil,et al. The TechQA Dataset , 2019, ACL.
[14] Wei-Cheng Chang,et al. Pre-training Tasks for Embedding-based Large-scale Retrieval , 2020, ICLR.
[15] W. Bruce Croft,et al. Open-Retrieval Conversational Question Answering , 2020, SIGIR.
[16] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[17] Quoc V. Le,et al. QANet: Combining Local Convolution with Global Self-Attention for Reading Comprehension , 2018, ICLR.
[18] Ming-Wei Chang,et al. REALM: Retrieval-Augmented Language Model Pre-Training , 2020, ICML.
[19] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[20] Ming Zhou,et al. Reinforced Mnemonic Reader for Machine Reading Comprehension , 2017, IJCAI.
[21] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[22] Thomas Wolf,et al. Transfer Learning in Natural Language Processing , 2019, NAACL.
[23] Yang Deng,et al. Knowledge as A Bridge: Improving Cross-domain Answer Selection with External Knowledge , 2018, COLING.
[24] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[25] Ming-Wei Chang,et al. Latent Retrieval for Weakly Supervised Open Domain Question Answering , 2019, ACL.
[26] Qiang Yang,et al. A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[27] Hannaneh Hajishirzi,et al. Question Answering through Transfer Learning from Large Fine-grained Supervision Data , 2017, ACL.
[28] Qingkai Zeng,et al. Crossing Variational Autoencoders for Answer Retrieval , 2020, ACL.
[29] Nitesh V. Chawla,et al. Graph Few-shot Learning via Knowledge Transfer , 2020, AAAI.
[30] Jimmy J. Lin,et al. End-to-End Open-Domain Question Answering with BERTserini , 2019, NAACL.
[31] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[32] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[33] Mark Andrew Greenwood,et al. Open-domain question answering , 2005 .
[34] Doug Downey,et al. Don’t Stop Pretraining: Adapt Language Models to Domains and Tasks , 2020, ACL.
[35] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[36] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[37] Percy Liang,et al. Know What You Don’t Know: Unanswerable Questions for SQuAD , 2018, ACL.