An Augmented Transformer Architecture for Natural Language Generation Tasks
暂无分享,去创建一个
Zhibin Lei | Wenye Li | Yang Liu | Hailiang Li | Du Tang | Adele Y. C. Wang | Wenye Li | Hailiang Li | Zhibin Lei | Adele Y. C. Wang | Yang Liu | Du Tang
[1] Christopher D. Manning,et al. Get To The Point: Summarization with Pointer-Generator Networks , 2017, ACL.
[2] Alon Lavie,et al. METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments , 2005, IEEvaluation@ACL.
[3] Dat Quoc Nguyen,et al. A Novel Neural Network Model for Joint POS Tagging and Graph-based Dependency Parsing , 2017, CoNLL.
[4] Alexander M. Rush,et al. OpenNMT: Open-Source Toolkit for Neural Machine Translation , 2017, ACL.
[5] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[6] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[7] Yoshua Bengio,et al. Gradient Flow in Recurrent Nets: the Difficulty of Learning Long-Term Dependencies , 2001 .
[8] Luowei Zhou,et al. End-to-End Dense Video Captioning with Masked Transformer , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[9] Slav Petrov,et al. Improved Transition-Based Parsing and Tagging with Neural Networks , 2015, EMNLP.
[10] Zhongfei Zhang,et al. TVT: Two-View Transformer Network for Video Captioning , 2018, ACML.
[11] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[12] Haizhou Li,et al. Joint Models for Chinese POS Tagging and Dependency Parsing , 2011, EMNLP.
[13] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[14] Daniel Jurafsky,et al. Sharp Nearby, Fuzzy Far Away: How Neural Language Models Use Context , 2018, ACL.
[15] Yen-Chun Chen,et al. Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting , 2018, ACL.
[16] Luke S. Zettlemoyer,et al. AllenNLP: A Deep Semantic Natural Language Processing Platform , 2018, ArXiv.
[17] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[18] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[19] Omer Levy,et al. Recurrent Additive Networks , 2017, ArXiv.
[20] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Jan Niehues,et al. Exploiting Linguistic Resources for Neural Machine Translation Using Multi-task Learning , 2017, WMT.
[22] Ophir Frieder,et al. On understanding and classifying web queries , 2006 .
[23] Weicun Zhang,et al. Boosted Transformer for Image Captioning , 2019, Applied Sciences.
[24] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[25] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[26] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[27] Rico Sennrich,et al. Linguistic Input Features Improve Neural Machine Translation , 2016, WMT.
[28] Joakim Nivre,et al. A Transition-Based System for Joint Part-of-Speech Tagging and Labeled Non-Projective Dependency Parsing , 2012, EMNLP.
[29] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[30] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[31] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[32] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[33] Ashish Vaswani,et al. Self-Attention with Relative Position Representations , 2018, NAACL.
[34] Ronald J. Williams,et al. A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.
[35] Mirella Lapata,et al. Long Short-Term Memory-Networks for Machine Reading , 2016, EMNLP.
[36] Lixiang Li,et al. Captioning Transformer with Stacked Attention Modules , 2018 .