暂无分享,去创建一个
Siu Cheung Hui | Yi Tay | Luu Anh Tuan | Yi Tay | Anh Tuan Luu | S. Hui
[1] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[2] Yelong Shen,et al. ReasoNet: Learning to Stop Reading in Machine Comprehension , 2016, CoCo@NIPS.
[3] Ting Liu,et al. Attention-over-Attention Neural Networks for Reading Comprehension , 2016, ACL.
[4] Hwee Tou Ng,et al. A Question-Focused Multi-Factor Attention Network for Question Answering , 2018, AAAI.
[5] Thomas S. Huang,et al. Dilated Recurrent Neural Networks , 2017, NIPS.
[6] Yu Zhang,et al. Simple Recurrent Units for Highly Parallelizable Recurrence , 2017, EMNLP.
[7] Richard Socher,et al. Dynamic Coattention Networks For Question Answering , 2016, ICLR.
[8] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[9] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[10] Rudolf Kadlec,et al. Text Understanding with the Attention Sum Reader Network , 2016, ACL.
[11] Shuohang Wang,et al. A Compare-Aggregate Model for Matching Text Sequences , 2016, ICLR.
[12] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[13] Yu Zhang,et al. Training RNNs as Fast as CNNs , 2017, EMNLP 2018.
[14] Yuxing Peng,et al. Mnemonic Reader for Machine Comprehension , 2017, ArXiv.
[15] Chris Dyer,et al. The NarrativeQA Reading Comprehension Challenge , 2017, TACL.
[16] Sebastian Riedel,et al. Constructing Datasets for Multi-hop Reading Comprehension Across Documents , 2017, TACL.
[17] Alessandro Moschitti,et al. Learning to Rank Short Text Pairs with Convolutional Deep Neural Networks , 2015, SIGIR.
[18] Guokun Lai,et al. RACE: Large-scale ReAding Comprehension Dataset From Examinations , 2017, EMNLP.
[19] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[20] Siu Cheung Hui,et al. Learning to Rank Question Answer Pairs with Holographic Dual LSTM Architecture , 2017, SIGIR.
[21] Alex Graves,et al. Neural Machine Translation in Linear Time , 2016, ArXiv.
[22] Richard Socher,et al. Quasi-Recurrent Neural Networks , 2016, ICLR.
[23] Wei Zhang,et al. R3: Reinforced Reader-Ranker for Open-Domain Question Answering , 2017, ArXiv.
[24] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[25] Ming Zhou,et al. Reinforced Mnemonic Reader for Machine Reading Comprehension , 2017, IJCAI.
[26] Kyunghyun Cho,et al. SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine , 2017, ArXiv.
[27] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[28] Zhiguo Wang,et al. Bilateral Multi-Perspective Matching for Natural Language Sentences , 2017, IJCAI.
[29] Xiaodong Liu,et al. Towards Human-level Machine Reading Comprehension: Reasoning and Inference with Multiple Strategies , 2017, ArXiv.
[30] Danqi Chen,et al. A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task , 2016, ACL.
[31] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[32] Siu Cheung Hui,et al. Cross Temporal Recurrent Networks for Ranking Question Answer Pairs , 2017, AAAI.
[33] Shuohang Wang,et al. Machine Comprehension Using Match-LSTM and Answer Pointer , 2016, ICLR.
[34] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[35] Ruslan Salakhutdinov,et al. Gated-Attention Readers for Text Comprehension , 2016, ACL.
[36] Mitesh M. Khapra,et al. ElimiNet: A Model for Eliminating Options for Reading Comprehension with Multiple Choice Questions , 2018, IJCAI.
[37] Ming Zhou,et al. Gated Self-Matching Networks for Reading Comprehension and Question Answering , 2017, ACL.