暂无分享,去创建一个
[1] Alexander M. Rush,et al. Character-Aware Neural Language Models , 2015, AAAI.
[2] Paul Over,et al. DUC in context , 2007, Inf. Process. Manag..
[3] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[4] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[5] Taku Kudo,et al. Subword Regularization: Improving Neural Network Translation Models with Multiple Subword Candidates , 2018, ACL.
[6] Masaaki Nagata,et al. Learning Compact Neural Word Embeddings by Parameter Space Sharing , 2016, IJCAI.
[7] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[8] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[9] Myle Ott,et al. Scaling Neural Machine Translation , 2018, WMT.
[10] Masaaki Nagata,et al. Character n-gram Embeddings to Improve RNN Language Models , 2019, AAAI.
[11] Naoaki Okazaki,et al. Positional Encoding to Control Output Sequence Length , 2019, NAACL.
[12] Yizhou Sun,et al. Learning K-way D-dimensional Discrete Code For Compact Embedding Representations , 2017, ICML.
[13] Ehud Rivlin,et al. Placing search in context: the concept revisited , 2002, TOIS.
[14] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[15] John B. Goodenough,et al. Contextual correlates of synonymy , 1965, CACM.
[16] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[17] Alexander M. Rush,et al. Sequence-Level Knowledge Distillation , 2016, EMNLP.
[18] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[19] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[20] Ming Yang,et al. Compressing Deep Convolutional Networks using Vector Quantization , 2014, ArXiv.
[21] Jason Weston,et al. A Neural Attention Model for Abstractive Sentence Summarization , 2015, EMNLP.
[22] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[23] Shuang Xu,et al. Towards Compact and Fast Neural Machine Translation Using a Combined Method , 2017, EMNLP.
[24] Myle Ott,et al. fairseq: A Fast, Extensible Toolkit for Sequence Modeling , 2019, NAACL.
[25] Hideki Nakayama,et al. Compressing Word Embeddings via Deep Compositional Code Learning , 2017, ICLR.
[26] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[27] Felix Hill,et al. SimLex-999: Evaluating Semantic Models With (Genuine) Similarity Estimation , 2014, CL.
[28] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[29] Ole Winther,et al. Hash Embeddings for Efficient Word Representations , 2017, NIPS.
[30] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[31] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[32] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[33] Masaaki Nagata,et al. Cutting-off Redundant Repeating Generations for Neural Abstractive Summarization , 2016, EACL.
[34] Yixin Chen,et al. Compressing Neural Networks with the Hashing Trick , 2015, ICML.
[35] Matt Post,et al. A Call for Clarity in Reporting BLEU Scores , 2018, WMT.
[36] Hannaneh Hajishirzi,et al. DeFINE: DEep Factorized INput Word Embeddings for Neural Sequence Modeling , 2020, ICLR.
[37] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[38] Kevin Gimpel,et al. Charagram: Embedding Words and Sentences via Character n-grams , 2016, EMNLP.