暂无分享,去创建一个
[1] Kevin Gimpel,et al. Towards Universal Paraphrastic Sentence Embeddings , 2015, ICLR.
[2] Gerard de Melo,et al. Exploring Semantic Properties of Sentence Embeddings , 2018, ACL.
[3] Samuel R. Bowman,et al. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference , 2017, NAACL.
[4] Christopher Joseph Pal,et al. Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning , 2018, ICLR.
[5] Felix Hill,et al. Learning Distributed Representations of Sentences from Unlabelled Data , 2016, NAACL.
[6] Chee Kheong Siew,et al. Extreme learning machine: Theory and applications , 2006, Neurocomputing.
[7] Yonatan Belinkov,et al. Fine-grained Analysis of Sentence Embeddings Using Auxiliary Prediction Tasks , 2016, ICLR.
[8] Peter Ford Dominey,et al. On-Line Processing of Grammatical Structure Using Reservoir Computing , 2012, ICANN.
[9] Omer Levy,et al. GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding , 2018, BlackboxNLP@EMNLP.
[10] B. Eynard,et al. Random matrices. , 2015, 1510.04430.
[11] Christopher D. Manning,et al. Baselines and Bigrams: Simple, Good Sentiment and Topic Classification , 2012, ACL.
[12] Quoc V. Le,et al. Distributed Representations of Sentences and Documents , 2014, ICML.
[13] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[14] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[15] Eric B. Baum,et al. On the capabilities of multilayer perceptrons , 1988, J. Complex..
[16] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[17] Holger Schwenk,et al. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data , 2017, EMNLP.
[18] Noah D. Goodman,et al. DisSent: Sentence Representation Learning from Explicit Discourse Relations , 2017, ArXiv.
[19] Stefan L. Frank,et al. Learn more by training less: systematicity in sentence processing by recurrent networks , 2006, Connect. Sci..
[20] Santosh S. Vempala,et al. The Random Projection Method , 2005, DIMACS Series in Discrete Mathematics and Theoretical Computer Science.
[21] Benjamin Recht,et al. Random Features for Large-Scale Kernel Machines , 2007, NIPS.
[22] Allyson Ettinger,et al. Probing for semantic evidence of composition by means of simple classification tasks , 2016, RepEval@ACL.
[23] Angeliki Lazaridou,et al. Jointly optimizing word representations for lexical and sentential tasks with the C-PHRASE model , 2015, ACL.
[24] Douwe Kiela,et al. SentEval: An Evaluation Toolkit for Universal Sentence Representations , 2018, LREC.
[25] Herbert Jaeger,et al. The''echo state''approach to analysing and training recurrent neural networks , 2001 .
[26] Samuel R. Bowman,et al. Discourse-Based Objectives for Fast Unsupervised Sentence Representation Learning , 2017, ArXiv.
[27] AI Koan,et al. Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning , 2008, NIPS.
[28] Willem H. Zuidema,et al. Visualisation and 'diagnostic classifiers' reveal how recurrent and recursive neural networks process hierarchical structure , 2017, J. Artif. Intell. Res..
[29] Allan Jabri,et al. Learning Visually Grounded Sentence Representations , 2018, NAACL.
[30] Rico Sennrich,et al. How Grammatical is Character-level Neural Machine Translation? Assessing MT Quality with Contrastive Translation Pairs , 2016, EACL.
[31] A. Gamba,et al. An outline of a mathematical theory of PAPA , 1961 .
[32] Yonatan Belinkov,et al. Understanding and Improving Morphological Learning in the Neural Machine Translation Decoder , 2017, IJCNLP.
[33] Jeffrey Pennington,et al. Dynamic Pooling and Unfolding Recursive Autoencoders for Paraphrase Detection , 2011, NIPS.
[34] Yann LeCun,et al. What is the best multi-stage architecture for object recognition? , 2009, 2009 IEEE 12th International Conference on Computer Vision.
[35] Zachary C. Lipton,et al. Troubling Trends in Machine Learning Scholarship , 2018, ACM Queue.
[36] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[37] Herbert Jaeger,et al. Reservoir computing approaches to recurrent neural network training , 2009, Comput. Sci. Rev..
[38] Matteo Pagliardini,et al. Unsupervised Learning of Sentence Embeddings Using Compositional n-Gram Features , 2017, NAACL.
[39] Dejan J. Sobajic,et al. Learning and generalization characteristics of the random vector Functional-link net , 1994, Neurocomputing.
[40] Marvin Minsky,et al. Perceptrons: An Introduction to Computational Geometry , 1969 .
[41] Thomas M. Cover,et al. Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition , 1965, IEEE Trans. Electron. Comput..
[42] Guillaume Lample,et al. What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties , 2018, ACL.
[43] Andrea Vedaldi,et al. Deep Image Prior , 2017, International Journal of Computer Vision.
[44] Le Song,et al. Scalable Kernel Methods via Doubly Stochastic Gradients , 2014, NIPS.
[45] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[46] Stefan L. Frank. Strong Systematicity in Sentence Processing by an Echo State Network , 2006, ICANN.
[47] Sanjeev Arora,et al. A Simple but Tough-to-Beat Baseline for Sentence Embeddings , 2017, ICLR.
[48] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[49] Kevin Gimpel,et al. Pushing the Limits of Paraphrastic Sentence Embeddings with Millions of Machine Translations , 2017, ArXiv.
[50] Guoyin Wang,et al. Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms , 2018, ACL.
[51] A. Gamba,et al. Further experiments with PAPA , 1961 .
[52] Robert P. W. Duin,et al. Feedforward neural networks with random weights , 1992, Proceedings., 11th IAPR International Conference on Pattern Recognition. Vol.II. Conference B: Pattern Recognition Methodology and Systems.
[53] Garrison W. Cottrell,et al. 2007 Special Issue: Learning grammatical structure with Echo State Networks , 2007 .
[54] Matthieu Geist,et al. Model-free POMDP optimisation of tutoring systems with echo-state networks , 2013, SIGDIAL Conference.
[55] Nan Hua,et al. Universal Sentence Encoder , 2018, ArXiv.
[56] Yonatan Belinkov,et al. What do Neural Machine Translation Models Learn about Morphology? , 2017, ACL.
[57] Marco Baroni,et al. Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks , 2017, ICML.
[58] Zhenghao Chen,et al. On Random Weights and Unsupervised Feature Learning , 2011, ICML.
[59] Emmanuel Dupoux,et al. Assessing the Ability of LSTMs to Learn Syntax-Sensitive Dependencies , 2016, TACL.
[60] Sanja Fidler,et al. Skip-Thought Vectors , 2015, NIPS.