暂无分享,去创建一个
[1] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[2] Jonathan Evans. Heuristic and analytic processes in reasoning , 1984 .
[3] Ruslan Salakhutdinov,et al. Neural Models for Reasoning over Multiple Mentions Using Coreference , 2018, NAACL.
[4] Hongzi Mao,et al. Interpreting Deep Learning-Based Networking Systems , 2019, SIGCOMM.
[5] Wei Shi,et al. Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification , 2016, ACL.
[6] Kai Liu,et al. Multi-Passage Machine Reading Comprehension with Cross-Passage Answer Verification , 2018, ACL.
[7] Bowen Zhou,et al. Multi-hop Reading Comprehension across Multiple Documents by Reasoning over Heterogeneous Graphs , 2019, ACL.
[8] Wilhelm Freiherr von Humboldt,et al. On Language: On the Diversity of Human Language Construction and Its Influence on the Mental Development of the Human Species , 2001 .
[9] Nicola De Cao,et al. Question Answering by Reasoning Across Documents with Graph Convolutional Networks , 2018, NAACL.
[10] Yu Cao,et al. BAG: Bi-directional Attention Entity Graph Convolutional Network for Multi-hop Reasoning Question Answering , 2019, NAACL.
[11] Razvan Pascanu,et al. Relational inductive biases, deep learning, and graph networks , 2018, ArXiv.
[12] Yoshimasa Tsuruoka,et al. A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks , 2016, EMNLP.
[13] Sebastian Riedel,et al. Constructing Datasets for Multi-hop Reading Comprehension Across Documents , 2017, TACL.
[14] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[15] Ting Liu,et al. Attention-over-Attention Neural Networks for Reading Comprehension , 2016, ACL.
[16] S. Sloman. The empirical case for two systems of reasoning. , 1996 .
[17] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[18] Samuel S. Schoenholz,et al. Neural Message Passing for Quantum Chemistry , 2017, ICML.
[19] Jason Weston,et al. The Goldilocks Principle: Reading Children's Books with Explicit Memory Representations , 2015, ICLR.
[20] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[21] S. Dehaene. Reading in the Brain: The New Science of How We Read , 2009 .
[22] Richard Socher,et al. Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering , 2019, ICLR.
[23] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[24] Aric Hagberg,et al. Exploring Network Structure, Dynamics, and Function using NetworkX , 2008, Proceedings of the Python in Science Conference.
[25] Ali Farhadi,et al. Bidirectional Attention Flow for Machine Comprehension , 2016, ICLR.
[26] Steven Bird,et al. NLTK: The Natural Language Toolkit , 2002, ACL.
[27] Max Welling,et al. Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.
[28] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[29] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[30] Chang Zhou,et al. Cognitive Graph for Multi-Hop Reading Comprehension at Scale , 2019, ACL.
[31] Yue Zhang,et al. Exploring Graph-structured Passage Representation for Multi-hop Reading Comprehension with Graph Neural Networks , 2018, ArXiv.
[32] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[33] Pietro Liò,et al. Graph Attention Networks , 2017, ICLR.
[34] Jan Eric Lenssen,et al. Fast Graph Representation Learning with PyTorch Geometric , 2019, ArXiv.