暂无分享,去创建一个
Razvan Pascanu | Yoshua Bengio | Kyunghyun Cho | Çaglar Gülçehre | Yoshua Bengio | Kyunghyun Cho | Çaglar Gülçehre | Razvan Pascanu
[1] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[2] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[3] Jürgen Schmidhuber,et al. Learning Complex, Extended Sequences Using the Principle of History Compression , 1992, Neural Computation.
[4] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[5] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[6] Yoshua Bengio,et al. Hierarchical Recurrent Neural Networks for Long-Term Dependencies , 1995, NIPS.
[7] Juha Karhunen,et al. An Unsupervised Ensemble Learning Method for Nonlinear Dynamic State-Space Models , 2002, Neural Computation.
[8] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[9] Yoshua. Bengio,et al. Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..
[10] Dieter Fox,et al. GP-BayesFilters: Bayesian filtering using Gaussian process prediction and observation models , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[11] Herbert Jaeger,et al. Discovering multiscale dynamical features with hierarchical Echo State Networks , 2008 .
[12] Quoc V. Le,et al. Measuring Invariances in Deep Networks , 2009, NIPS.
[13] J. Schmidhuber,et al. A Novel Connectionist System for Unconstrained Handwriting Recognition , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[14] James Martens,et al. Deep learning via Hessian-free optimization , 2010, ICML.
[15] Nicolas Le Roux,et al. Deep Belief Networks Are Compact Universal Approximators , 2010, Neural Computation.
[16] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[17] Lukás Burget,et al. Extensions of recurrent neural network language model , 2011, 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[18] Ilya Sutskever,et al. Learning Recurrent Neural Networks with Hessian-Free Optimization , 2011, ICML.
[19] Ilya Sutskever,et al. SUBWORD LANGUAGE MODELING WITH NEURAL NETWORKS , 2011 .
[20] Hugo Larochelle,et al. The Neural Autoregressive Distribution Estimator , 2011, AISTATS.
[21] Yoshua Bengio,et al. Shallow vs. Deep Sum-Product Networks , 2011, NIPS.
[22] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[23] Yoshua Bengio,et al. Deep Sparse Rectifier Neural Networks , 2011, AISTATS.
[24] Yoshua Bengio,et al. Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.
[25] Geoffrey E. Hinton,et al. Generating Text with Recurrent Neural Networks , 2011, ICML.
[26] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[27] Yoshua Bengio,et al. Modeling Temporal Dependencies in High-Dimensional Sequences: Application to Polyphonic Music Generation and Transcription , 2012, ICML.
[28] Razvan Pascanu,et al. Theano: new features and speed improvements , 2012, ArXiv.
[29] Vysoké Učení,et al. Statistical Language Models Based on Neural Networks , 2012 .
[30] Tapani Raiko,et al. Deep Learning Made Easier by Linear Transformations in Perceptrons , 2012, AISTATS.
[31] Tara N. Sainath,et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition , 2012 .
[32] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[33] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[34] Ronan Collobert,et al. Recurrent Convolutional Neural Networks for Scene Parsing , 2013, ArXiv.
[35] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[36] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[37] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[38] Jianshu Chen,et al. A Primal-Dual Method for Training Recurrent Neural Networks Constrained by the Echo-State Property , 2013 .
[39] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[40] Benjamin Schrauwen,et al. Training and Analysing Deep Recurrent Neural Networks , 2013, NIPS.
[41] Yoshua Bengio,et al. Maxout Networks , 2013, ICML.
[42] Li Deng,et al. A New Method for Learning Deep Recurrent Neural Networks , 2013, ICLR.
[43] Yoshua Bengio,et al. Better Mixing via Deep Representations , 2012, ICML.
[44] Ronan Collobert,et al. Recurrent Convolutional Neural Networks for Scene Labeling , 2014, ICML.
[45] Razvan Pascanu,et al. Learned-Norm Pooling for Deep Feedforward and Recurrent Neural Networks , 2013, ECML/PKDD.
[46] Christian Osendorfer,et al. On Fast Dropout and its Applicability to Recurrent Networks , 2013, ICLR.
[47] Razvan Pascanu,et al. On the number of inference regions of deep feed forward networks with piece-wise linear activations , 2013, ICLR.
[48] Razvan Pascanu,et al. Revisiting Natural Gradient for Deep Networks , 2013, ICLR.