Combining Global Models for Parsing Universal Dependencies
暂无分享,去创建一个
Yao Cheng | Tianze Shi | Xilun Chen | Felix G. Wu | Felix Wu | Tianze Shi | Xilun Chen | Yao Cheng
[1] Benno Stein,et al. Improving the Reproducibility of PAN's Shared Tasks: - Plagiarism Detection, Author Identification, and Author Profiling , 2014, CLEF.
[2] Giorgio Satta,et al. Dynamic Programming Algorithms for Transition-Based Dependency Parsers , 2011, ACL.
[3] Kenji Sagae,et al. Dynamic Programming for Linear-Time Incremental Parsing , 2010, ACL.
[4] Alon Lavie,et al. Parser Combination by Reparsing , 2006, NAACL.
[5] Noah A. Smith,et al. Improved Transition-based Parsing by Modeling Characters instead of Words with LSTMs , 2015, EMNLP.
[6] Yuan Zhang,et al. Stack-propagation: Improved Representation Learning for Syntax , 2016, ACL.
[7] Sampo Pyysalo,et al. Universal Dependencies v1: A Multilingual Treebank Collection , 2016, LREC.
[8] Timothy Dozat,et al. Deep Biaffine Attention for Neural Dependency Parsing , 2016, ICLR.
[9] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[10] Giorgio Satta,et al. Efficient Parsing for Bilexical Context-Free Grammars and Head Automaton Grammars , 1999, ACL.
[11] Jason Eisner,et al. Three New Probabilistic Models for Dependency Parsing: An Exploration , 1996, COLING.
[12] Jan Hajic,et al. UDPipe: Trainable Pipeline for Processing CoNLL-U Files Performing Tokenization, Morphological Analysis, POS Tagging and Parsing , 2016, LREC.
[13] Jürgen Schmidhuber,et al. Framewise phoneme classification with bidirectional LSTM and other neural network architectures , 2005, Neural Networks.
[14] Maria Leonor Pacheco,et al. of the Association for Computational Linguistics: , 2001 .
[15] Ji Ma,et al. SyntaxNet Models for the CoNLL 2017 Shared Task , 2017, ArXiv.
[16] Nizar Habash,et al. CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies , 2017, CoNLL.
[17] Eliyahu Kiperwasser,et al. Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations , 2016, TACL.
[18] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[19] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[20] Martin Potthast,et al. CoNLL 2018 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies , 2018, CoNLL.
[21] Surya Ganguli,et al. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.
[22] Ben Taskar,et al. Learning structured prediction models: a large margin approach , 2005, ICML.
[23] Barbara Plank,et al. Multilingual Part-of-Speech Tagging with Bidirectional Long Short-Term Memory Models and Auxiliary Loss , 2016, ACL.
[24] Sean R Eddy,et al. What is dynamic programming? , 2004, Nature Biotechnology.
[25] Yuji Matsumoto,et al. Universal Dependencies 2.0 – CoNLL 2017 Shared Task Development and Test Data , 2017 .
[26] Noah A. Smith. Linguistic Structure Prediction , 2011, Synthesis Lectures on Human Language Technologies.
[27] James Cross,et al. Incremental Parsing with Minimal Features Using Bi-Directional LSTM , 2016, ACL.
[28] Kevin Duh,et al. DyNet: The Dynamic Neural Network Toolkit , 2017, ArXiv.
[29] Philipp Koehn,et al. Synthesis Lectures on Human Language Technologies , 2016 .
[30] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.