Text Understanding with a Hybrid Neural Network Based Learning
暂无分享,去创建一个
Kai Gao | Hua-Ping Zhang | Shen Gao | Kai Gao | Hua-Ping Zhang | Shen Gao
[1] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[2] Rada Mihalcea,et al. TextRank: Bringing Order into Text , 2004, EMNLP.
[3] Hayato Kobayashi,et al. Summarization Based on Embedding Distributions , 2015, EMNLP.
[4] Mirella Lapata,et al. Neural Summarization by Extracting Sentences and Words , 2016, ACL.
[5] Chin-Yew Lin,et al. ROUGE: A Package for Automatic Evaluation of Summaries , 2004, ACL 2004.
[6] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[7] J. Schmidhuber,et al. Framewise phoneme classification with bidirectional LSTM networks , 2005, Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005..
[8] Kuldip K. Paliwal,et al. Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..
[9] Phil Blunsom,et al. Teaching Machines to Read and Comprehend , 2015, NIPS.
[10] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[11] Oren Etzioni,et al. Machine Reading at the University of Washington , 2010, HLT-NAACL 2010.
[12] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[13] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[14] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[15] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[16] Lukasz Kaiser,et al. Sentence Compression by Deletion with LSTMs , 2015, EMNLP.
[17] Ellen Riloff,et al. A Rule-based Question Answering System for Reading Comprehension Tests , 2000 .
[18] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[19] Konstantin Lopyrev,et al. Generating News Headlines with Recurrent Neural Networks , 2015, ArXiv.
[20] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[21] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.