暂无分享,去创建一个
John Blitzer | Kilian Q. Weinberger | Ni Lao | Felix Wu | Guandao Yang | Felix Wu | John Blitzer | N. Lao | Guandao Yang
[1] Quoc V. Le,et al. Semi-supervised Sequence Learning , 2015, NIPS.
[2] Yuxing Peng,et al. Reinforced Mnemonic Reader for Machine Comprehension , 2017 .
[3] Deng Cai,et al. Smarnet: Teaching Machines to Read and Comprehend Like Human , 2017, ArXiv.
[4] Alex Graves,et al. Neural Machine Translation in Linear Time , 2016, ArXiv.
[5] Andrew Chou,et al. Semantic Parsing on Freebase from Question-Answer Pairs , 2013, EMNLP.
[6] Jason Weston,et al. Key-Value Memory Networks for Directly Reading Documents , 2016, EMNLP.
[7] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[8] Petr Baudis,et al. Modeling of the Question Answering Task in the YodaQA System , 2015, CLEF.
[9] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[10] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Eunsol Choi,et al. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension , 2017, ACL.
[13] Ming Zhou,et al. Gated Self-Matching Networks for Reading Comprehension and Question Answering , 2017, ACL.
[14] Vladlen Koltun,et al. Multi-Scale Context Aggregation by Dilated Convolutions , 2015, ICLR.
[15] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[16] Martín Abadi,et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.
[17] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[18] Christopher Clark,et al. Simple and Effective Multi-Paragraph Reading Comprehension , 2017, ACL.
[19] Yann Dauphin,et al. Language Modeling with Gated Convolutional Networks , 2016, ICML.
[20] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[21] Samy Bengio,et al. Can Active Memory Replace Attention? , 2016, NIPS.
[22] Jakob Uszkoreit,et al. A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.
[23] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[24] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[25] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[26] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[27] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[28] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.