暂无分享,去创建一个
[1] John D. Hunter,et al. Matplotlib: A 2D Graphics Environment , 2007, Computing in Science & Engineering.
[2] Yoshua Bengio,et al. Unitary Evolution Recurrent Neural Networks , 2015, ICML.
[3] Deep Ray,et al. Deep learning observables in computational fluid dynamics , 2019, J. Comput. Phys..
[4] Vladlen Koltun,et al. Trellis Networks for Sequence Modeling , 2018, ICLR.
[5] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[6] Lovekesh Vig,et al. Long Short Term Memory Networks for Anomaly Detection in Time Series , 2015, ESANN.
[7] Yasuhiro Fujiwara,et al. Preventing Gradient Explosions in Gated Recurrent Units , 2017, NIPS.
[8] Eldad Haber,et al. Stable architectures for deep neural networks , 2017, ArXiv.
[9] Fei-Fei Li,et al. Deep visual-semantic alignments for generating image descriptions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[11] Geoffrey E. Hinton,et al. A Simple Way to Initialize Recurrent Networks of Rectified Linear Units , 2015, ArXiv.
[12] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[13] Yuanzhi Li,et al. Can SGD Learn Recurrent Neural Networks with Provable Generalization? , 2019, NeurIPS.
[14] E. Hairer,et al. Geometric Numerical Integration: Structure Preserving Algorithms for Ordinary Differential Equations , 2004 .
[15] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[16] Ed H. Chi,et al. AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks , 2019, ICLR.
[17] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[18] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[19] Ole Winther,et al. Sequential Neural Models with Stochastic Layers , 2016, NIPS.
[20] Kunihiko Fukushima,et al. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position , 1980, Biological Cybernetics.
[21] Yoshua Bengio,et al. A Recurrent Latent Variable Model for Sequential Data , 2015, NIPS.
[22] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[23] Gaël Varoquaux,et al. The NumPy Array: A Structure for Efficient Numerical Computation , 2011, Computing in Science & Engineering.
[24] Razvan Pascanu,et al. Understanding the exploding gradient problem , 2012, ArXiv.
[25] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[26] Jun Liang,et al. Residual Recurrent Neural Networks for Learning Sequential Representations , 2018, Inf..
[27] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[28] Paul J. Werbos,et al. Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.
[29] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[30] Alex Graves,et al. Sequence Transduction with Recurrent Neural Networks , 2012, ArXiv.
[31] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[32] Feng Liu,et al. On Computation and Generalization of Generative Adversarial Networks under Spectrum Control , 2019, ICLR.
[33] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.