Summarizing Articles into Sentences by Hierarchical Attention Model and RNN Language Model
暂无分享,去创建一个
[1] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[2] Jürgen Schmidhuber,et al. Learning to forget: continual prediction with LSTM , 1999 .
[3] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[4] Diyi Yang,et al. Hierarchical Attention Networks for Document Classification , 2016, NAACL.
[5] Alexander M. Rush,et al. Abstractive Sentence Summarization with Attentive Recurrent Neural Networks , 2016, NAACL.
[6] Shingo Mabu,et al. A Sentence Summarizer using Recurrent Neural Network and Attention-Based Encoder , 2017 .
[7] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[8] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[9] Ronald J. Williams,et al. Gradient-based learning algorithms for recurrent networks and their computational complexity , 1995 .
[10] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[11] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[12] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[13] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[14] Chengqi Zhang,et al. Bi-Directional Block Self-Attention for Fast and Memory-Efficient Sequence Modeling , 2018, ICLR.
[15] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[16] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.