An Empirical Exploration of Recurrent Network Architectures
暂无分享,去创建一个
Wojciech Zaremba | Ilya Sutskever | Rafal Józefowicz | Ilya Sutskever | Wojciech Zaremba | R. Józefowicz | I. Sutskever
[1] Daniel J. Amit,et al. Modeling brain function: the world of attractor neural networks, 1st Edition , 1989 .
[2] Sepp Hochreiter,et al. Untersuchungen zu dynamischen neuronalen Netzen , 1991 .
[3] Geoffrey E. Hinton,et al. Lesioning an attractor network: investigations of acquired dyslexia , 1991 .
[4] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[5] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[6] David C. Plaut,et al. Semantic and Associative Priming in a Distributed Attractor Network , 1995 .
[7] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[8] Jürgen Schmidhuber,et al. Learning to forget: continual prediction with LSTM , 1999 .
[9] Herbert Jaeger,et al. The''echo state''approach to analysing and training recurrent neural networks , 2001 .
[10] Harald Haas,et al. Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication , 2004, Science.
[11] Julian Togelius,et al. Evolving Memory Cell Structures for Sequence Learning , 2009, ICANN.
[12] James Martens,et al. Deep learning via Hessian-free optimization , 2010, ICML.
[13] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[14] Ilya Sutskever,et al. Learning Recurrent Neural Networks with Hessian-Free Optimization , 2011, ICML.
[15] Clément Farabet,et al. Torch7: A Matlab-like Environment for Machine Learning , 2011, NIPS 2011.
[16] Yoshua Bengio,et al. Modeling Temporal Dependencies in High-Dimensional Sequences: Application to Polyphonic Music Generation and Transcription , 2012, ICML.
[17] Vysoké Učení,et al. Statistical Language Models Based on Neural Networks , 2012 .
[18] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[19] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[20] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[21] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[22] Wojciech Zaremba,et al. Learning to Execute , 2014, ArXiv.
[23] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[24] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[25] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[26] Marc'Aurelio Ranzato,et al. Learning Longer Memory in Recurrent Neural Networks , 2014, ICLR.
[27] Jürgen Schmidhuber,et al. LSTM: A Search Space Odyssey , 2015, IEEE Transactions on Neural Networks and Learning Systems.