暂无分享,去创建一个
[1] Zhiyuan Liu,et al. A C-LSTM Neural Network for Text Classification , 2015, ArXiv.
[2] Dan Roth,et al. Learning Question Classifiers , 2002, COLING.
[3] Yue Zhang,et al. A Search-Based Dynamic Reranking Model for Dependency Parsing , 2016, ACL.
[4] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[5] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[6] Jason Weston,et al. A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.
[7] Eduard H. Hovy,et al. When Are Tree Structures Necessary for Deep Learning of Representations? , 2015, EMNLP.
[8] Jürgen Schmidhuber,et al. Recurrent nets that time and count , 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks. IJCNN 2000. Neural Computing: New Challenges and Perspectives for the New Millennium.
[9] Noah A. Smith,et al. Transition-Based Dependency Parsing with Stack Long Short-Term Memory , 2015, ACL.
[10] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[11] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[12] Noah A. Smith,et al. Recurrent Neural Network Grammars , 2016, NAACL.
[13] Dan Klein,et al. Accurate Unlexicalized Parsing , 2003, ACL.
[14] James R. Curran,et al. Parsing the WSJ Using CCG and Log-Linear Models , 2004, ACL.
[15] Jeffrey L. Elman,et al. Finding Structure in Time , 1990, Cogn. Sci..
[16] Andrew Y. Ng,et al. Parsing with Compositional Vector Grammars , 2013, ACL.
[17] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[18] Navdeep Jaitly,et al. Hybrid speech recognition with Deep Bidirectional LSTM , 2013, 2013 IEEE Workshop on Automatic Speech Recognition and Understanding.
[19] Makoto Miwa,et al. End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures , 2016, ACL.
[20] Michael Collins,et al. Head-Driven Statistical Models for Natural Language Parsing , 2003, CL.
[21] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[22] Hongyu Guo,et al. Long Short-Term Memory Over Tree Structures , 2015, ArXiv.
[23] Geoffrey E. Hinton,et al. Grammar as a Foreign Language , 2014, NIPS.
[24] Christopher D. Manning,et al. Global Belief Recursive Neural Networks , 2014, NIPS.
[25] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[26] Phong Le,et al. Compositional Distributional Semantics with Long Short Term Memory , 2015, *SEMEVAL.
[27] Luísa Coheur,et al. From symbolic to sub-symbolic information in question classification , 2011, Artificial Intelligence Review.
[28] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[29] Liang Lu,et al. Tree Recurrent Neural Networks with Application to Language Modeling , 2015, ArXiv.
[30] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[31] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[32] Xuanjing Huang,et al. A Re-ranking Model for Dependency Parser with Recursive Convolutional Neural Network , 2015, ACL.
[33] Yue Zhang,et al. Fast and Accurate Shift-Reduce Constituent Parsing , 2013, ACL.
[34] Eugene Charniak,et al. Parsing as Language Modeling , 2016, EMNLP.
[35] Andrew Y. Ng,et al. Parsing Natural Scenes and Natural Language with Recursive Neural Networks , 2011, ICML.