暂无分享,去创建一个
Yoshimasa Tsuruoka | Richard Socher | Caiming Xiong | Kazuma Hashimoto | R. Socher | Caiming Xiong | Kazuma Hashimoto | Yoshimasa Tsuruoka
[1] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[2] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[3] Andrew Y. Ng,et al. Semantic Compositionality through Recursive Matrix-Vector Spaces , 2012, EMNLP.
[4] Noah A. Smith,et al. Transition-Based Dependency Parsing with Stack Long Short-Term Memory , 2015, ACL.
[5] Yan Pan,et al. Modelling Sentence Pairs with Tree-structured Attentive Encoder , 2016, COLING.
[6] Bowen Zhou,et al. ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs , 2015, TACL.
[7] Jun Zhao,et al. Recurrent Convolutional Neural Networks for Text Classification , 2015, AAAI.
[8] Noah A. Smith,et al. What Do Recurrent Neural Network Grammars Learn About Syntax? , 2016, EACL.
[9] Martial Hebert,et al. Cross-Stitch Networks for Multi-task Learning , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Eliyahu Kiperwasser,et al. Easy-First Dependency Parsing with Hierarchical Tree LSTMs , 2016, TACL.
[11] Christopher Kermorvant,et al. Dropout Improves Recurrent Neural Networks for Handwriting Recognition , 2013, 2014 14th International Conference on Frontiers in Handwriting Recognition.
[12] Yuan Zhang,et al. Stack-propagation: Improved Representation Learning for Syntax , 2016, ACL.
[13] Slav Petrov,et al. Improved Transition-Based Parsing and Tagging with Neural Networks , 2015, EMNLP.
[14] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[15] Yusuke Miyao,et al. Learning with Lookahead: Can History-Based Models Rival Globally Optimized Models? , 2011, CoNLL.
[16] Kevin Gimpel,et al. Charagram: Embedding Words and Sentences via Character n-grams , 2016, EMNLP.
[17] Jason Eisner. Efficient Normal-Form Parsing for Combinatory Categorial Grammar , 1996, ACL.
[18] Sebastian Riedel,et al. Deep Semi-Supervised Learning with Linguistically Motivated Sequence Labeling Task Hierarchies , 2016, ArXiv.
[19] Yuji Matsumoto,et al. Chunking with Support Vector Machines , 2001, NAACL.
[20] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[21] Eugene Charniak,et al. Parsing as Language Modeling , 2016, EMNLP.
[22] Richard Socher,et al. Ask Me Anything: Dynamic Memory Networks for Natural Language Processing , 2015, ICML.
[23] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[24] Mirella Lapata,et al. Dependency Parsing as Head Selection , 2016, EACL.
[25] Yoshimasa Tsuruoka,et al. Neural Machine Translation with Source-Side Latent Graph Parsing , 2017, EMNLP.
[26] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[27] Jürgen Schmidhuber,et al. Framewise phoneme classification with bidirectional LSTM and other neural network architectures , 2005, Neural Networks.
[28] Anders Søgaard,et al. Deep multi-task learning with low level tasks supervised at lower layers , 2016, ACL.
[29] Slav Petrov,et al. Globally Normalized Transition-Based Neural Networks , 2016, ACL.
[30] Slav Petrov,et al. Structured Training for Neural Network Transition-Based Parsing , 2015, ACL.
[31] Dan Klein,et al. Feature-Rich Part-of-Speech Tagging with a Cyclic Dependency Network , 2003, NAACL.
[32] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[33] Jun Suzuki,et al. Semi-Supervised Sequential Labeling and Segmentation Using Giga-Word Scale Unlabeled Data , 2008, ACL.
[34] Giuseppe Attardi,et al. Chunking and Dependency Parsing , 2008 .
[35] Wang Ling,et al. Finding Function in Form: Compositional Character Models for Open Vocabulary Word Representation , 2015, EMNLP.
[36] Makoto Miwa,et al. End-to-End Relation Extraction using LSTMs on Sequences and Tree Structures , 2016, ACL.
[37] Bernd Bohnet,et al. Top Accuracy and Fast Dependency Parsing is not a Contradiction , 2010, COLING.
[38] Anders Søgaard,et al. Semi-supervised condensed nearest neighbor for part-of-speech tagging , 2011, ACL.
[39] Eduard H. Hovy,et al. End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF , 2016, ACL.
[40] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[41] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[42] Quoc V. Le,et al. Multi-task Sequence to Sequence Learning , 2015, ICLR.
[43] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[44] Xinyun Chen. Under Review as a Conference Paper at Iclr 2017 Delving into Transferable Adversarial Ex- Amples and Black-box Attacks , 2016 .
[45] Yoshimasa Tsuruoka,et al. Tree-to-Sequence Attentional Neural Machine Translation , 2016, ACL.
[46] Zhen-Hua Ling,et al. Enhancing and Combining Sequential and Tree LSTM for Natural Language Inference , 2016, ArXiv.
[47] Derek Hoiem,et al. Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[48] M. Marelli,et al. SemEval-2014 Task 1: Evaluation of Compositional Distributional Semantic Models on Full Sentences through Semantic Relatedness and Textual Entailment , 2014, *SEMEVAL.
[49] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[50] Alice Lai,et al. Illinois-LH: A Denotational and Distributional Approach to Semantics , 2014, *SEMEVAL.
[51] Razvan Pascanu,et al. Progressive Neural Networks , 2016, ArXiv.
[52] Kyunghyun Cho,et al. Gated Word-Character Recurrent Language Model , 2016, EMNLP.
[53] Yoshua Bengio,et al. Maxout Networks , 2013, ICML.
[54] Makoto Miwa,et al. Word Embedding-based Antonym Detection using Thesauri and Distributional Information , 2015, NAACL.