Enhancing Sentence Embedding with Generalized Pooling
暂无分享,去创建一个
[1] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[2] Rui Yan,et al. Natural Language Inference by Tree-Based Convolution and Heuristic Matching , 2015, ACL.
[3] Hong Yu,et al. Neural Tree Indexers for Text Understanding , 2016, EACL.
[4] Mohit Bansal,et al. Shortcut-Stacked Sentence Encoders for Multi-Domain Inference , 2017, RepEval@EMNLP.
[5] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[6] Hong Yu,et al. Neural Semantic Encoders , 2016, EACL.
[7] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[8] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[9] Jeffrey Pennington,et al. Semi-Supervised Recursive Autoencoders for Predicting Sentiment Distributions , 2011, EMNLP.
[10] Quoc V. Le,et al. Distributed Representations of Sentences and Documents , 2014, ICML.
[11] Li-Rong Dai,et al. Exploring Question Understanding and Adaptation in Neural-Network-Based Question Answering , 2017, ArXiv.
[12] Siu Cheung Hui,et al. A Compare-Propagate Architecture with Alignment Factorization for Natural Language Inference , 2017, ArXiv.
[13] Yang Liu,et al. Learning Natural Language Inference using Bidirectional LSTM model and Inner-Attention , 2016, ArXiv.
[14] Jihun Choi,et al. Unsupervised Learning of Task-Specific Tree Structures with Tree-LSTMs , 2017, ArXiv.
[15] Mirella Lapata,et al. Long Short-Term Memory-Networks for Machine Reading , 2016, EMNLP.
[16] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[17] Felix Hill,et al. Learning Distributed Representations of Sentences from Unlabelled Data , 2016, NAACL.
[18] Hongyu Guo,et al. Long Short-Term Memory Over Recursive Structures , 2015, ICML.
[19] Zhen-Hua Ling,et al. Enhanced LSTM for Natural Language Inference , 2016, ACL.
[20] Phil Blunsom,et al. Reasoning about Entailment with Neural Attention , 2015, ICLR.
[21] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[22] Sanja Fidler,et al. Order-Embeddings of Images and Language , 2015, ICLR.
[23] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[24] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[25] John Salvatier,et al. Theano: A Python framework for fast computation of mathematical expressions , 2016, ArXiv.
[26] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[27] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[28] Zhen-Hua Ling,et al. Distraction-based neural networks for modeling documents , 2016, IJCAI 2016.
[29] Zhen-Hua Ling,et al. Neural Natural Language Inference Models Enhanced with External Knowledge , 2017, ACL.
[30] Christopher Potts,et al. A Fast Unified Model for Parsing and Sentence Understanding , 2016, ACL.
[31] Zhen-Hua Ling,et al. Recurrent Neural Network-Based Sentence Encoder with Gated Attention for Natural Language Inference , 2017, RepEval@EMNLP.
[32] Zhen-Hua Ling,et al. Distraction-Based Neural Networks for Modeling Document , 2016, IJCAI.
[33] Tao Shen,et al. DiSAN: Directional Self-Attention Network for RNN/CNN-free Language Understanding , 2017, AAAI.
[34] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[35] Phil Blunsom,et al. A Convolutional Neural Network for Modelling Sentences , 2014, ACL.
[36] Chengqi Zhang,et al. Reinforced Self-Attention Network: a Hybrid of Hard and Soft Attention for Sequence Modeling , 2018, IJCAI.
[37] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[38] Jakob Uszkoreit,et al. A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.
[39] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[40] Yoshua Bengio,et al. Fine-grained attention mechanism for neural machine translation , 2018, Neurocomputing.