Wider and Deeper, Cheaper and Faster: Tensorized LSTMs for Sequence Learning
暂无分享,去创建一个
Liang Xiao | David Barber | Daxue Liu | Zhen He | Hangen He | Shaobing Gao | D. Barber | Daxue Liu | Liang Xiao | Hangen He | Zhen He | Shaobing Gao
[1] Yu Zhang,et al. Training RNNs as Fast as CNNs , 2017, EMNLP 2018.
[2] Yoshua Bengio,et al. Architectural Complexity Measures of Recurrent Neural Networks , 2016, NIPS.
[3] Luc Van Gool,et al. Dynamic Filter Networks , 2016, NIPS.
[4] Jürgen Schmidhuber,et al. Parallel Multi-Dimensional LSTM, With Application to Fast Biomedical Volumetric Image Segmentation , 2015, NIPS.
[5] Misha Denil,et al. Predicting Parameters in Deep Learning , 2014 .
[6] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[7] Yu Zhang,et al. Simple Recurrent Units for Highly Parallelizable Recurrence , 2017, EMNLP.
[8] Geoffrey E. Hinton,et al. A Simple Way to Initialize Recurrent Networks of Rectified Linear Units , 2015, ArXiv.
[9] Alexander Novikov,et al. Tensorizing Neural Networks , 2015, NIPS.
[10] Jürgen Schmidhuber,et al. Learning to Forget: Continual Prediction with LSTM , 2000, Neural Computation.
[11] Viorica Patraucean,et al. Spatio-temporal video autoencoder with differentiable memory , 2015, ArXiv.
[12] Lin Yang,et al. Combining Fully Convolutional and Recurrent Neural Networks for 3D Biomedical Image Segmentation , 2016, NIPS.
[13] Yoshua Bengio,et al. Unitary Evolution Recurrent Neural Networks , 2015, ICML.
[14] Yoshua. Bengio,et al. Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..
[15] Claire Cardie,et al. Modeling Compositionality with Multiplicative Recurrent Neural Networks , 2014, ICLR.
[16] Wojciech Zaremba,et al. An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.
[17] Quoc V. Le,et al. HyperNetworks , 2016, ICLR.
[18] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[19] Geoffrey E. Hinton,et al. Generating Text with Recurrent Neural Networks , 2011, ICML.
[20] Lukasz Kaiser,et al. Neural GPUs Learn Algorithms , 2015, ICLR.
[21] Yoshua Bengio,et al. Hierarchical Multiscale Recurrent Neural Networks , 2016, ICLR.
[22] Ying Zhang,et al. On Multiplicative Integration with Recurrent Neural Networks , 2016, NIPS.
[23] Clément Farabet,et al. Torch7: A Matlab-like Environment for Machine Learning , 2011, NIPS 2011.
[24] Thomas S. Huang,et al. Dilated Recurrent Neural Networks , 2017, NIPS.
[25] Alexander Novikov,et al. Ultimate tensorization: compressing convolutional and FC layers alike , 2016, ArXiv.
[26] Alex Graves,et al. Grid Long Short-Term Memory , 2015, ICLR.
[27] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[28] Steve Renals,et al. Multiplicative LSTM for sequence modelling , 2016, ICLR.
[29] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[30] Aaron C. Courville,et al. Recurrent Batch Normalization , 2016, ICLR.
[31] Tobias Grüning,et al. Cells in Multidimensional Recurrent Neural Networks , 2016, J. Mach. Learn. Res..
[32] Geoffrey E. Hinton,et al. Speech recognition with deep recurrent neural networks , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[33] Angelika Steger,et al. Fast-Slow Recurrent Neural Networks , 2017, NIPS.
[34] Philip H. S. Torr,et al. Recurrent Instance Segmentation , 2015, ECCV.
[35] Lawrence D. Jackel,et al. Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.
[36] Jeffrey L. Elman,et al. Finding Structure in Time , 1990, Cogn. Sci..
[37] Richard Socher,et al. Quasi-Recurrent Neural Networks , 2016, ICLR.
[38] Phil Blunsom,et al. Optimizing Performance of Recurrent Neural Networks on GPUs , 2016, ArXiv.
[39] Geoffrey E. Hinton,et al. Factored conditional restricted Boltzmann Machines for modeling motion style , 2009, ICML '09.
[40] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[41] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[42] Jürgen Schmidhuber,et al. Recurrent Highway Networks , 2016, ICML.
[43] Yoshua Bengio,et al. Gated Feedback Recurrent Neural Networks , 2015, ICML.
[44] Luca Bertinetto,et al. Learning feed-forward one-shot learners , 2016, NIPS.
[45] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[46] Dit-Yan Yeung,et al. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting , 2015, NIPS.
[47] Lin Wu,et al. Deep Recurrent Convolutional Networks for Video-based Person Re-identification: An End-to-End Approach , 2016, ArXiv.
[48] Les E. Atlas,et al. Full-Capacity Unitary Recurrent Neural Networks , 2016, NIPS.
[49] Alex Graves,et al. Adaptive Computation Time for Recurrent Neural Networks , 2016, ArXiv.
[50] Jürgen Schmidhuber,et al. Learning to Control Fast-Weight Memories: An Alternative to Dynamic Recurrent Networks , 1992, Neural Computation.
[51] Samy Bengio,et al. Can Active Memory Replace Attention? , 2016, NIPS.
[52] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[53] Erich Elsen,et al. Persistent RNNs: Stashing Recurrent Weights On-Chip , 2016, ICML.