暂无分享,去创建一个
Yan Xu | Farhad Bin Siddique | Pascale Fung | Dan Su | Elham J. Barezi | Tiezheng Yu | Pascale Fung | Dan Su | Yan Xu | Tiezheng Yu
[1] Hannaneh Hajishirzi,et al. Multi-hop Reading Comprehension through Question Decomposition and Rescoring , 2019, ACL.
[2] Yan Xu,et al. Generalizing Question Answering System with Pre-trained Language Model Fine-tuning , 2019, EMNLP.
[3] Shashi Narayan,et al. Split and Rephrase , 2017, EMNLP.
[4] Omer Levy,et al. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension , 2019, ACL.
[5] Jonghyun Choi,et al. Are You Smarter Than a Sixth Grader? Textbook Question Answering for Multimodal Machine Comprehension , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Shao Hui Huang,et al. What We Know So Far (As of March 26, 2020) About COVID-19—An MRT Point of View , 2020, Journal of Medical Imaging and Radiation Sciences.
[7] Kazuhiro Takemoto,et al. Global COVID-19 transmission rate is influenced by precipitation seasonality and the speed of climate temperature warming , 2020, medRxiv.
[8] Axel-Cyrille Ngonga Ngomo,et al. BioASQ: A Challenge on Large-Scale Biomedical Semantic Indexing and Question Answering , 2012, AAAI Fall Symposium: Information Retrieval and Knowledge Discovery in Biomedical Text.
[9] Yoshua Bengio,et al. HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering , 2018, EMNLP.
[10] Jaewoo Kang,et al. BioBERT: a pre-trained biomedical language representation model for biomedical text mining , 2019, Bioinform..
[11] Daphne Stannard. COVID-19: Impact on Perianesthesia Nursing Areas , 2020, Journal of PeriAnesthesia Nursing.
[12] Kyunghyun Cho,et al. Unsupervised Question Decomposition for Question Answering , 2020, EMNLP.
[13] Xiaodong Liu,et al. Unified Language Model Pre-training for Natural Language Understanding and Generation , 2019, NeurIPS.
[14] Philip Bachman,et al. NewsQA: A Machine Comprehension Dataset , 2016, Rep4NLP@ACL.
[15] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[16] Peter Clark,et al. Modeling Biological Processes for Reading Comprehension , 2014, EMNLP.
[17] Jimmy J. Lin,et al. Anserini , 2018, Journal of Data and Information Quality.
[18] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[19] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[20] Kevin Gimpel,et al. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations , 2019, ICLR.
[21] Duangdao Wichadakul,et al. Extractive Text Summarization Using Ontology and Graph-Based Method , 2019, 2019 IEEE 4th International Conference on Computer and Communication Systems (ICCCS).
[22] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[23] Gabriel Stanovsky,et al. DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs , 2019, NAACL.
[24] Ming-Wei Chang,et al. Natural Questions: A Benchmark for Question Answering Research , 2019, TACL.
[25] Mirella Lapata,et al. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization , 2018, EMNLP.
[26] Jimmy J. Lin,et al. Rapidly Bootstrapping a Question Answering Dataset for COVID-19 , 2020, ArXiv.
[27] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[28] Kyunghyun Cho,et al. SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine , 2017, ArXiv.