暂无分享,去创建一个
[1] Daniel Marcu,et al. Learning as search optimization: approximate large margin methods for structured prediction , 2005, ICML.
[2] Slav Petrov,et al. Globally Normalized Transition-Based Neural Networks , 2016, ACL.
[3] John Langford,et al. Search-based structured prediction , 2009, Machine Learning.
[4] Erik F. Tjong Kim Sang,et al. Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition , 2003, CoNLL.
[5] Mark Steedman,et al. CCGbank: A Corpus of CCG Derivations and Dependency Structures Extracted from the Penn Treebank , 2007, CL.
[6] Joakim Nivre,et al. Training Deterministic Parsers with Non-Deterministic Oracles , 2013, TACL.
[7] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[8] Ashish Vaswani,et al. Supertagging With LSTMs , 2016, NAACL.
[9] Alexander M. Rush,et al. Sequence-to-Sequence Learning as Beam-Search Optimization , 2016, EMNLP.
[10] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[11] Gabriel Synnaeve,et al. A Fully Differentiable Beam Search Decoder , 2019, ICML.
[12] Sabine Buchholz,et al. Introduction to the CoNLL-2000 Shared Task Chunking , 2000, CoNLL/LLL.
[13] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[14] Stephen Clark,et al. A Tale of Two Parsers: Investigating and Combining Graph-based and Transition-based Dependency Parsing , 2008, EMNLP.
[15] Yang Guo,et al. Structured Perceptron with Inexact Search , 2012, HLT-NAACL.
[16] Frank Hutter,et al. SGDR: Stochastic Gradient Descent with Warm Restarts , 2016, ICLR.
[17] Graham Neubig,et al. A Continuous Relaxation of Beam Search for End-to-end Training of Neural Sequence Models , 2017, AAAI.
[18] J. Andrew Bagnell,et al. Reinforcement and Imitation Learning via Interactive No-Regret Learning , 2014, ArXiv.
[19] Joakim Nivre,et al. A Dynamic Oracle for Arc-Eager Dependency Parsing , 2012, COLING.
[20] J. Andrew Bagnell,et al. Efficient Reductions for Imitation Learning , 2010, AISTATS.
[21] Chris Dyer,et al. An Empirical Investigation of Global and Local Normalization for Recurrent Neural Sequence Models Using a Continuous Relaxation to Beam Search , 2019, NAACL.
[22] Noah A. Smith,et al. Training with Exploration Improves a Greedy Stack LSTM Parser , 2016, EMNLP.
[23] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[24] Brian Roark,et al. Incremental Parsing with the Perceptron Algorithm , 2004, ACL.
[25] Samy Bengio,et al. Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.
[26] John Langford,et al. Learning to Search Better than Your Teacher , 2015, ICML.
[27] Geoffrey J. Gordon,et al. Learning Beam Search Policies via Imitation Learning , 2018, NeurIPS.
[28] Shinji Watanabe,et al. Promising Accurate Prefix Boosting for Sequence-to-sequence ASR , 2019, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[29] Marc'Aurelio Ranzato,et al. Classical Structured Prediction Losses for Sequence to Sequence Learning , 2017, NAACL.