Translation Quality Estimation using Recurrent Neural Network
暂无分享,去创建一个
[1] Lucia Specia,et al. Data enhancement and selection strategies for the word-level Quality Estimation , 2015, WMT@EMNLP.
[2] Yoshua Bengio,et al. Hierarchical Probabilistic Neural Network Language Model , 2005, AISTATS.
[3] Philipp Koehn,et al. Findings of the 2013 Workshop on Statistical Machine Translation , 2013, WMT@ACL.
[4] Philipp Koehn,et al. Findings of the 2015 Workshop on Statistical Machine Translation , 2015, WMT@EMNLP.
[5] Philipp Koehn,et al. Findings of the 2012 Workshop on Statistical Machine Translation , 2012, WMT@NAACL-HLT.
[6] Dongfeng Cai,et al. Strategy-Based Technology for Estimating MT Quality , 2015, WMT@EMNLP.
[7] Philipp Koehn,et al. Findings of the 2014 Workshop on Statistical Machine Translation , 2014, WMT@ACL.
[8] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[9] Andrew Y. Ng,et al. Parsing with Compositional Vector Grammars , 2013, ACL.
[10] Geoffrey Zweig,et al. Recurrent neural networks for language understanding , 2013, INTERSPEECH.
[11] Lucia Specia,et al. SHEF-NN: Translation Quality Estimation with Neural Networks , 2015, WMT@EMNLP.
[12] Matteo Negri,et al. FBK-UPV-UEdin participation in the WMT14 Quality Estimation shared-task , 2014, WMT@ACL.
[13] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[14] Geoffrey E. Hinton,et al. A Scalable Hierarchical Distributed Language Model , 2008, NIPS.
[15] Nello Cristianini,et al. Estimating the Sentence-Level Quality of Machine Translation Systems , 2009, EAMT.
[16] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[17] Matthew D. Zeiler. ADADELTA: An Adaptive Learning Rate Method , 2012, ArXiv.
[18] Stefan Riezler,et al. QUality Estimation from ScraTCH (QUETCH): Deep Learning for Word-level Translation Quality Estimation , 2015, WMT@EMNLP.
[19] Razvan Pascanu,et al. Theano: new features and speed improvements , 2012, ArXiv.
[20] Jean-Luc Gauvain,et al. Connectionist language modeling for large vocabulary continuous speech recognition , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.
[21] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[22] Alex Kulesza,et al. Confidence Estimation for Machine Translation , 2004, COLING.
[23] Geoffrey Zweig,et al. Spoken language understanding using long short-term memory neural networks , 2014, 2014 IEEE Spoken Language Technology Workshop (SLT).
[24] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[25] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[26] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[27] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[28] Philipp Koehn,et al. Europarl: A Parallel Corpus for Statistical Machine Translation , 2005, MTSUMMIT.
[29] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[30] Quoc V. Le,et al. Exploiting Similarities among Languages for Machine Translation , 2013, ArXiv.
[31] Holger Schwenk,et al. Continuous space language models , 2007, Comput. Speech Lang..
[32] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[33] Paul J. Werbos,et al. Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.