暂无分享,去创建一个
Qichen Li | Jun Xia | Jian Li | Xiaoke Jiang
[1] Myle Ott,et al. Scaling Neural Machine Translation , 2018, WMT.
[2] Kyunghyun Cho,et al. Dynamic Meta-Embeddings for Improved Sentence Representations , 2018, EMNLP.
[3] Yuan Luo,et al. Graph Convolutional Networks for Text Classification , 2018, AAAI.
[4] Dilin Wang,et al. Improving Neural Language Modeling via Adversarial Training , 2019, ICML.
[5] Wenpeng Yin,et al. Learning Word Meta-Embeddings , 2016, ACL.
[6] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[8] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[9] Guoyin Wang,et al. Joint Embedding of Words and Labels for Text Classification , 2018, ACL.
[10] Tomas Mikolov,et al. Bag of Tricks for Efficient Text Classification , 2016, EACL.
[11] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[12] Yuhong Guo,et al. Time-aware Large Kernel Convolutions , 2020, ICML.
[13] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[14] Qiaozhu Mei,et al. PTE: Predictive Text Embedding through Large-scale Heterogeneous Text Networks , 2015, KDD.
[15] Richard Socher,et al. Weighted Transformer Network for Machine Translation , 2017, ArXiv.
[16] Xavier Bresson,et al. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering , 2016, NIPS.
[17] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[18] Ankur Bapna,et al. The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation , 2018, ACL.
[19] Rico Sennrich,et al. Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures , 2018, EMNLP.
[20] Danushka Bollegala,et al. Frustratingly Easy Meta-Embedding - Computing Meta-Embeddings by Averaging Source Word Embeddings , 2018, NAACL-HLT.
[21] Quoc V. Le,et al. Distributed Representations of Sentences and Documents , 2014, ICML.
[22] Xuanjing Huang,et al. Recurrent Neural Network for Text Classification with Multi-Task Learning , 2016, IJCAI.
[23] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[24] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[25] Ken-ichi Kawarabayashi,et al. Think Globally, Embed Locally - Locally Linear Meta-embedding of Words , 2018, IJCAI.
[26] Ilya Sutskever,et al. Generating Long Sequences with Sparse Transformers , 2019, ArXiv.
[27] Joan Bruna,et al. Spectral Networks and Locally Connected Networks on Graphs , 2013, ICLR.
[28] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[29] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[30] Sven Laur,et al. Linear Ensembles of Word Embedding Models , 2017, NODALIDA.
[31] Yann Dauphin,et al. Pay Less Attention with Lightweight and Dynamic Convolutions , 2019, ICLR.
[32] Lior Wolf,et al. Using the Output Embedding to Improve Language Models , 2016, EACL.
[33] Edouard Grave,et al. Adaptive Attention Span in Transformers , 2019, ACL.
[34] Ondrej Bojar,et al. An Exploration of Word Embedding Initialization in Deep-Learning Tasks , 2017, ICON.
[35] Eneko Agirre,et al. Uncovering Divergent Linguistic Information in Word Embeddings with Lessons for Intrinsic and Extrinsic Evaluation , 2018, CoNLL.
[36] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[37] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[38] Danushka Bollegala,et al. Angular-Based Word Meta-Embedding Learning , 2018, ArXiv.
[39] Yiming Yang,et al. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context , 2019, ACL.
[40] Quoc V. Le,et al. Massive Exploration of Neural Machine Translation Architectures , 2017, EMNLP.
[41] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[42] Ashish Vaswani,et al. Self-Attention with Relative Position Representations , 2018, NAACL.
[43] Guoyin Wang,et al. Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms , 2018, ACL.
[44] Noah A. Smith,et al. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) , 2016, ACL 2016.