暂无分享,去创建一个
[1] Zhiyong Luo,et al. Combination of Convolutional and Recurrent Neural Network for Sentiment Analysis of Short Texts , 2016, COLING.
[2] Tao Wang,et al. Convolutional Neural Networks over Tree Structures for Programming Language Processing , 2014, AAAI.
[3] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[4] Quoc V. Le,et al. Attention Augmented Convolutional Networks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[5] Tong Zhang,et al. Modeling Localness for Self-Attention Networks , 2018, EMNLP.
[6] Xing Wang,et al. Multi-Granularity Self-Attention for Neural Machine Translation , 2019, EMNLP.
[7] Ting Liu,et al. Gaussian Transformer: A Lightweight Approach for Natural Language Inference , 2019, AAAI.
[8] Wei Wu,et al. Phrase-level Self-Attention Networks for Universal Sentence Encoding , 2018, EMNLP.
[9] Ankur Bapna,et al. The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation , 2018, ACL.
[10] Rico Sennrich,et al. Neural Machine Translation of Rare Words with Subword Units , 2015, ACL.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Zhaopeng Tu,et al. Convolutional Self-Attention Networks , 2019, NAACL.
[13] Salim Roukos,et al. Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.
[14] Xing Wang,et al. Self-Attention with Structural Position Representations , 2019, EMNLP.
[15] Xing Wang,et al. Context-Aware Self-Attention Networks , 2019, AAAI.
[16] Christopher Potts,et al. Tree-Structured Composition in Neural Networks without Tree-Structured Architectures , 2015, CoCo@NIPS.
[17] M. Tanenhaus. Afterword The impact of “The cognitive basis for linguistic structures” , 2013 .
[18] Tiejun Zhao,et al. Improving Neural Machine Translation with Neural Syntactic Distance , 2019, NAACL.
[19] Xing Wang,et al. Exploiting Sentential Context for Neural Machine Translation , 2019, ACL.
[20] Christof Monz,et al. The Importance of Being Recurrent for Modeling Hierarchical Structure , 2018, EMNLP.
[21] Guillaume Lample,et al. What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties , 2018, ACL.
[22] Noam Chomsky,et al. वाक्यविन्यास का सैद्धान्तिक पक्ष = Aspects of the theory of syntax , 1965 .
[23] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[24] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[25] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Edouard Grave,et al. Adaptive Attention Span in Transformers , 2019, ACL.
[27] Luke S. Zettlemoyer,et al. Transformers with convolutional context for ASR , 2019, ArXiv.
[28] Aaron C. Courville,et al. Ordered Neurons: Integrating Tree Structures into Recurrent Neural Networks , 2018, ICLR.
[29] Xing Wang,et al. Modeling Recurrence for Transformer , 2019, NAACL.
[30] Yidong Chen,et al. Deep Semantic Role Labeling with Self-Attention , 2017, AAAI.
[31] Christopher D. Manning,et al. Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks , 2015, ACL.
[32] Luke S. Zettlemoyer,et al. Deep Contextualized Word Representations , 2018, NAACL.
[33] Bowen Zhou,et al. A Structured Self-attentive Sentence Embedding , 2017, ICLR.
[34] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[35] J. Hayes. Cognition and the development of language , 1970 .
[36] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[37] Noam Chomsky,et al. Aspects of the Theory of Syntax , 1970 .
[38] Shuming Shi,et al. Exploiting Deep Representations for Neural Machine Translation , 2018, EMNLP.