暂无分享,去创建一个
[1] O. Marschall. Evaluating biological plausibility of learning algorithms the lazy way , 2019 .
[2] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[3] Timothy P Lillicrap,et al. Backpropagation through time and the brain , 2019, Current Opinion in Neurobiology.
[4] James Martens,et al. On the Variance of Unbiased Online Recurrent Optimization , 2019, ArXiv.
[5] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[6] Angelika Steger,et al. Optimal Kronecker-Sum Approximation of Real Time Recurrent Learning , 2019, ICML.
[7] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[8] Daniel Kifer,et al. Online Learning of Recurrent Neural Architectures by Locally Aligning Distributed Representations , 2018, ArXiv.
[9] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[10] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[11] Max Jaderberg,et al. Understanding Synthetic Gradients and Decoupled Neural Interfaces , 2017, ICML.
[12] James M Murray,et al. Local online learning in recurrent networks with random feedback , 2018, bioRxiv.
[13] Paul J. Werbos,et al. Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.
[14] Kyunghyun Cho,et al. Using local plasticity rules to train recurrent neural networks , 2019, ArXiv.
[15] Alex Graves,et al. Decoupled Neural Interfaces using Synthetic Gradients , 2016, ICML.
[16] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] David Reitter,et al. Learning to Adapt by Minimizing Discrepancy , 2017, ArXiv.
[18] Sergio Gomez Colmenarejo,et al. Hybrid computing using a neural network with dynamic external memory , 2016, Nature.
[19] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[20] Angelika Steger,et al. Approximating Real-Time Recurrent Learning with Random Kronecker Factors , 2018, NeurIPS.
[21] Colin J. Akerman,et al. Random synaptic feedback weights support error backpropagation for deep learning , 2016, Nature Communications.
[22] Yoshua Bengio,et al. Describing Multimedia Content Using Attention-Based Encoder-Decoder Networks , 2015, IEEE Transactions on Multimedia.
[23] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[24] Ronald J. Williams,et al. A Learning Algorithm for Continually Running Fully Recurrent Neural Networks , 1989, Neural Computation.
[25] Daniel Kifer,et al. Continual Learning of Recurrent Neural Networks by Locally Aligning Distributed Representations , 2018, IEEE Transactions on Neural Networks and Learning Systems.
[26] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[27] Herbert Jaeger,et al. Reservoir computing approaches to recurrent neural network training , 2009, Comput. Sci. Rev..
[28] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[29] Timothy P Lillicrap,et al. Towards deep learning with segregated dendrites , 2016, eLife.
[30] James M Murray,et al. Local online learning in recurrent networks with random feedback , 2018, bioRxiv.
[31] Yann Ollivier,et al. Unbiased Online Recurrent Optimization , 2017, ICLR.
[32] L. Deng,et al. The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.
[33] Christopher D. Manning,et al. Effective Approaches to Attention-based Neural Machine Translation , 2015, EMNLP.
[34] Yoshua Bengio,et al. Dendritic cortical microcircuits approximate the backpropagation algorithm , 2018, NeurIPS.
[35] Ingmar Kanitscheider,et al. Kernel RNN Learning (KeRNL) , 2018, ICLR.