暂无分享,去创建一个
Chengqi Zhang | Tao Shen | Tianyi Zhou | Guodong Long | Jing Jiang | Tao Shen | Tianyi Zhou | Guodong Long | Jing Jiang | Chengqi Zhang
[1] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[2] Rui Yan,et al. Natural Language Inference by Tree-Based Convolution and Heuristic Matching , 2015, ACL.
[3] Navdeep Jaitly,et al. Hybrid speech recognition with Deep Bidirectional LSTM , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.
[4] Tao Shen,et al. DiSAN: Directional Self-Attention Network for RNN/CNN-free Language Understanding , 2017, AAAI.
[5] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.
[6] Phil Blunsom,et al. A Convolutional Neural Network for Modelling Sentences , 2014, ACL.
[7] Malvina Nissim,et al. The Meaning Factory: Formal Semantics for Recognizing Textual Entailment and Determining Semantic Similarity , 2014, *SEMEVAL.
[8] Yu Zhang,et al. End-to-End Adversarial Memory Network for Cross-domain Sentiment Classification , 2017, IJCAI.
[9] Alexander M. Rush,et al. Structured Attention Networks , 2017, ICLR.
[10] Samuel R. Bowman,et al. Ruminating Reader: Reasoning with Gated Multi-hop Attention , 2017, QA@ACL.
[11] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[12] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[13] Hong Yu,et al. Neural Tree Indexers for Text Understanding , 2016, EACL.
[14] Yu Zhang,et al. Training RNNs as Fast as CNNs , 2017, EMNLP 2018.
[15] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[16] Claire Cardie,et al. Annotating Expressions of Opinions and Emotions in Language , 2005, Lang. Resour. Evaluation.
[17] Christopher Potts,et al. A Fast Unified Model for Parsing and Sentence Understanding , 2016, ACL.
[18] Hang Li,et al. Neural Responding Machine for Short-Text Conversation , 2015, ACL.
[19] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[20] Alexandros Potamianos,et al. Structural Attention Neural Networks for improved sentiment analysis , 2017, EACL.
[21] Dan Roth,et al. Learning Question Classifiers , 2002, COLING.
[22] Hong Yu,et al. Neural Semantic Encoders , 2016, EACL.
[23] Yang Liu,et al. Learning Natural Language Inference using Bidirectional LSTM model and Inner-Attention , 2016, ArXiv.
[24] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[25] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[26] Yue Zhang,et al. Bidirectional Tree-Structured LSTM with Head Lexicalization , 2016, ArXiv.
[27] Alex Graves,et al. Neural Machine Translation in Linear Time , 2016, ArXiv.
[28] Yann Dauphin,et al. Language Modeling with Gated Convolutional Networks , 2016, ICML.
[29] Yuxing Peng,et al. Reinforced Mnemonic Reader for Machine Comprehension , 2017 .
[30] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[31] Bo Pang,et al. Seeing Stars: Exploiting Class Relationships for Sentiment Categorization with Respect to Rating Scales , 2005, ACL.
[32] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Yoshua Bengio,et al. Deep Sparse Rectifier Neural Networks , 2011, AISTATS.
[34] Zhiguo Wang,et al. Multi-Perspective Context Matching for Machine Comprehension , 2016, ArXiv.
[35] Quoc V. Le,et al. Grounded Compositional Semantics for Finding and Describing Images with Sentences , 2014, TACL.
[36] Xiaoyan Zhu,et al. Linguistically Regularized LSTMs for Sentiment Classification , 2016, ArXiv.
[37] Ming Zhou,et al. Gated Self-Matching Networks for Reading Comprehension and Question Answering , 2017, ACL.
[38] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[39] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[40] Diyi Yang,et al. Hierarchical Attention Networks for Document Classification , 2016, NAACL.
[41] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[42] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[43] Han Zhao,et al. Self-Adaptive Hierarchical Sentence Model , 2015, IJCAI.
[44] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[45] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[46] Man Lan,et al. ECNU: One Stone Two Birds: Ensemble of Heterogenous Measures for Semantic Relatedness and Textual Entailment , 2014, *SEMEVAL.
[47] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[48] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[49] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[50] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[51] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[52] Yi Yang,et al. More is Less: A More Complicated Network with Less Inference Complexity , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[53] Bo Pang,et al. A Sentimental Education: Sentiment Analysis Using Subjectivity Summarization Based on Minimum Cuts , 2004, ACL.
[54] Zhen-Hua Ling,et al. Recurrent Neural Network-Based Sentence Encoder with Gated Attention for Natural Language Inference , 2017, RepEval@EMNLP.
[55] Jason Weston,et al. End-To-End Memory Networks , 2015, NIPS.
[56] Sanja Fidler,et al. Order-Embeddings of Images and Language , 2015, ICLR.
[57] Regina Barzilay,et al. Molding CNNs for text: non-linear, non-consecutive convolutions , 2015, EMNLP.
[58] Bing Liu,et al. Mining and summarizing customer reviews , 2004, KDD.