暂无分享,去创建一个
Jürgen Schmidhuber | Jan Koutník | Rupesh Kumar Srivastava | Julian G. Zilly | J. Schmidhuber | R. Srivastava | J. Koutník | J. Zilly
[1] S. Linnainmaa. Taylor expansion of the accumulated rounding error , 1976 .
[2] F. Kozin,et al. System modeling and optimization : proceedings of the 10th IFIP Conference, New York City, USA, August 31-September 4, 1981 , 1982 .
[3] F. Kozin,et al. System Modeling and Optimization , 1982 .
[4] Paul J. Werbos,et al. Applications of advances in nonlinear sensitivity analysis , 1982 .
[5] PAUL J. WERBOS,et al. Generalization of backpropagation with application to a recurrent gas market model , 1988, Neural Networks.
[6] Jürgen Schmidhuber,et al. Reinforcement Learning in Markovian and Non-Markovian Environments , 1990, NIPS.
[7] Sepp Hochreiter,et al. Untersuchungen zu dynamischen neuronalen Netzen , 1991 .
[8] Jürgen Schmidhuber,et al. Learning Complex, Extended Sequences Using the Principle of History Compression , 1992, Neural Computation.
[9] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[10] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[11] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[12] Jürgen Schmidhuber,et al. Learning to Forget: Continual Prediction with LSTM , 2000, Neural Computation.
[13] Yoshua Bengio,et al. Gradient Flow in Recurrent Nets: the Difficulty of Learning Long-Term Dependencies , 2001 .
[14] Jürgen Schmidhuber,et al. Sequence Labelling in Structured Domains with Hierarchical Recurrent Neural Networks , 2007, IJCAI.
[15] Yoshua Bengio,et al. Scaling learning algorithms towards AI , 2007 .
[16] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[17] Clément Farabet,et al. Torch7: A Matlab-like Environment for Machine Learning , 2011, NIPS 2011.
[18] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[19] Jürgen Schmidhuber,et al. Self-Delimiting Neural Networks , 2012, ArXiv.
[20] Yoshua Bengio,et al. Modeling Temporal Dependencies in High-Dimensional Sequences: Application to Polyphonic Music Generation and Transcription , 2012, ICML.
[21] Geoffrey Zweig,et al. Context dependent recurrent neural network language model , 2012, 2012 IEEE Spoken Language Technology Workshop (SLT).
[22] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[23] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[24] Jürgen Schmidhuber,et al. First Experiments with PowerPlay , 2012, Neural networks : the official journal of the International Neural Network Society.
[25] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[26] Franco Scarselli,et al. On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures , 2014, IEEE Transactions on Neural Networks and Learning Systems.
[27] Razvan Pascanu,et al. How to Construct Deep Recurrent Neural Networks , 2013, ICLR.
[28] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[29] Jürgen Schmidhuber,et al. Deep learning in neural networks: An overview , 2014, Neural Networks.
[30] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[31] Jürgen Schmidhuber,et al. Training Very Deep Networks , 2015, NIPS.
[32] Yoshua Bengio,et al. Gated Feedback Recurrent Neural Networks , 2015, ICML.
[33] Geoffrey E. Hinton,et al. A Simple Way to Initialize Recurrent Networks of Rectified Linear Units , 2015, ArXiv.
[34] Alex Graves,et al. Adaptive Computation Time for Recurrent Neural Networks , 2016, ArXiv.
[35] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[36] Ying Zhang,et al. On Multiplicative Integration with Recurrent Neural Networks , 2016, NIPS.
[37] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] Yonghui Wu,et al. Exploring the Limits of Language Modeling , 2016, ArXiv.
[39] Hakan Inan,et al. Improved Learning through Augmenting the Loss , 2016 .
[40] Alex Graves,et al. Grid Long Short-Term Memory , 2015, ICLR.
[41] Alexander M. Rush,et al. Character-Aware Neural Language Models , 2015, AAAI.
[42] Yu Zhang,et al. Highway long short-term memory RNNS for distant speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[43] Hakan Inan,et al. Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling , 2016, ICLR.
[44] Jürgen Schmidhuber,et al. Highway and Residual Networks learn Unrolled Iterative Estimation , 2016, ICLR.
[45] Steve Renals,et al. Multiplicative LSTM for sequence modelling , 2016, ICLR.
[46] Quoc V. Le,et al. HyperNetworks , 2016, ICLR.
[47] Lior Wolf,et al. Using the Output Embedding to Improve Language Models , 2016, EACL.
[48] Yoshua Bengio,et al. Hierarchical Multiscale Recurrent Neural Networks , 2016, ICLR.
[49] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[50] Aaron C. Courville,et al. Recurrent Batch Normalization , 2016, ICLR.
[51] Jürgen Schmidhuber,et al. LSTM: A Search Space Odyssey , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[52] Richard Socher,et al. Pointer Sentinel Mixture Models , 2016, ICLR.