暂无分享,去创建一个
Christopher Joseph Pal | Yoshua Bengio | Dmitriy Serdyuk | Alessandro Sordoni | Nan Rosemary Ke | Yoshua Bengio | C. Pal | Dmitriy Serdyuk | Alessandro Sordoni
[1] Yoshua Bengio,et al. End-to-end attention-based large vocabulary speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[2] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[3] Hugo Larochelle,et al. Neural Autoregressive Distribution Estimation , 2016, J. Mach. Learn. Res..
[4] Christoph Meinel,et al. Image Captioning with Deep Bidirectional LSTMs , 2016, ACM Multimedia.
[5] Ferenc Huszar,et al. How (not) to Train your Generative Model: Scheduled Sampling, Likelihood, Adversary? , 2015, ArXiv.
[6] Roland Memisevic,et al. Regularizing RNNs by Stabilizing Activations , 2015, ICLR.
[7] Rongrong Ji,et al. Asynchronous Bidirectional Decoding for Neural Machine Translation , 2018, AAAI.
[8] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[9] Samy Bengio,et al. Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks , 2015, NIPS.
[10] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[11] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[12] Yoshua Bengio,et al. Memory Augmented Neural Networks with Wormhole Connections , 2017, ArXiv.
[13] Kuldip K. Paliwal,et al. Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..
[14] Chris Dyer,et al. On the State of the Art of Evaluation in Neural Language Models , 2017, ICLR.
[15] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[16] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[17] Max Welling,et al. Markov Chain Monte Carlo and Variational Inference: Bridging the Gap , 2014, ICML.
[18] Yoshua Bengio,et al. Attention-Based Models for Speech Recognition , 2015, NIPS.
[19] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[20] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[21] Daniel Jurafsky,et al. Learning to Decode for Future Success , 2017, ArXiv.
[22] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[23] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[24] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[25] Chong Wang,et al. TopicRNN: A Recurrent Neural Network with Long-Range Semantic Dependency , 2016, ICLR.
[26] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[27] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[28] Sepp Hochreiter,et al. Untersuchungen zu dynamischen neuronalen Netzen , 1991 .
[29] Yoshua Bengio,et al. Task Loss Estimation for Sequence Prediction , 2015, ArXiv.
[30] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[31] Tao Mei,et al. Boosting Image Captioning with Attributes , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[32] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[33] Hugo Larochelle,et al. MADE: Masked Autoencoder for Distribution Estimation , 2015, ICML.
[34] Eduard H. Hovy,et al. End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF , 2016, ACL.
[35] Erhardt Barth,et al. Recurrent Dropout without Memory Loss , 2016, COLING.
[36] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[37] Yoshua Bengio,et al. Blocks and Fuel: Frameworks for deep learning , 2015, ArXiv.
[38] Tom Schaul,et al. The Predictron: End-To-End Learning and Planning , 2016, ICML.
[39] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[40] Yoshua Bengio,et al. Professor Forcing: A New Algorithm for Training Recurrent Networks , 2016, NIPS.
[41] Philip Bachman,et al. An Architecture for Deep, Hierarchical Generative Models , 2016, NIPS.
[42] Alex Graves,et al. Neural Turing Machines , 2014, ArXiv.
[43] Fei-Fei Li,et al. Deep visual-semantic alignments for generating image descriptions , 2015, CVPR.
[44] David Vázquez,et al. PixelVAE: A Latent Variable Model for Natural Images , 2016, ICLR.
[45] Daniel Povey,et al. The Kaldi Speech Recognition Toolkit , 2011 .
[46] Quoc V. Le,et al. Adding Gradient Noise Improves Learning for Very Deep Networks , 2015, ArXiv.
[47] Quoc V. Le,et al. Listen, Attend and Spell , 2015, ArXiv.
[48] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[49] Inchul Song,et al. RNNDROP: A novel dropout for RNNS in ASR , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).
[50] Joelle Pineau,et al. A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues , 2016, AAAI.
[51] Yoshua Bengio,et al. Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations , 2016, ICLR.
[52] Richard Socher,et al. Knowing When to Look: Adaptive Attention via a Visual Sentinel for Image Captioning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[53] Nicolas Usunier,et al. Improving Neural Language Models with a Continuous Cache , 2016, ICLR.
[54] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[55] Pieter Abbeel,et al. Value Iteration Networks , 2016, NIPS.
[56] John Salvatier,et al. Theano: A Python framework for fast computation of mathematical expressions , 2016, ArXiv.
[57] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.