Twin Networks: Matching the Future for Sequence Generation
暂无分享,去创建一个
Christopher Joseph Pal | Yoshua Bengio | Dmitriy Serdyuk | Adam Trischler | Alessandro Sordoni | Nan Rosemary Ke | Yoshua Bengio | C. Pal | Dmitriy Serdyuk | Adam Trischler | Alessandro Sordoni | A. Trischler
[1] Eduard H. Hovy,et al. End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF , 2016, ACL.
[2] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[3] Yoshua Bengio,et al. Attention-Based Models for Speech Recognition , 2015, NIPS.
[4] Tapani Raiko,et al. Iterative Neural Autoregressive Distribution Estimator NADE-k , 2014, NIPS.
[5] Erhardt Barth,et al. Recurrent Dropout without Memory Loss , 2016, COLING.
[6] Fei-Fei Li,et al. Deep visual-semantic alignments for generating image descriptions , 2015, CVPR.
[7] David Vázquez,et al. PixelVAE: A Latent Variable Model for Natural Images , 2016, ICLR.
[8] Pieter Abbeel,et al. Value Iteration Networks , 2016, NIPS.
[9] Daniel Povey,et al. The Kaldi Speech Recognition Toolkit , 2011 .
[10] Quoc V. Le,et al. Adding Gradient Noise Improves Learning for Very Deep Networks , 2015, ArXiv.
[11] Yoshua Bengio,et al. Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.
[12] Chong Wang,et al. TopicRNN: A Recurrent Neural Network with Long-Range Semantic Dependency , 2016, ICLR.
[13] Razvan Pascanu,et al. On the difficulty of training recurrent neural networks , 2012, ICML.
[14] Hugo Larochelle,et al. MADE: Masked Autoencoder for Distribution Estimation , 2015, ICML.
[15] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[16] Yoshua Bengio,et al. End-to-end attention-based large vocabulary speech recognition , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[17] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[18] Yoshua Bengio,et al. Blocks and Fuel: Frameworks for deep learning , 2015, ArXiv.
[19] Yoshua Bengio,et al. Memory Augmented Neural Networks with Wormhole Connections , 2017, ArXiv.
[20] Yoshua Bengio,et al. Professor Forcing: A New Algorithm for Training Recurrent Networks , 2016, NIPS.
[21] Philip Bachman,et al. An Architecture for Deep, Hierarchical Generative Models , 2016, NIPS.
[22] Alex Graves,et al. Neural Turing Machines , 2014, ArXiv.
[23] Quoc V. Le,et al. Listen, Attend and Spell , 2015, ArXiv.
[24] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[25] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[26] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[27] Joelle Pineau,et al. A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues , 2016, AAAI.
[28] Yoshua Bengio,et al. Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations , 2016, ICLR.
[29] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[30] Yoshua Bengio,et al. Unitary Evolution Recurrent Neural Networks , 2015, ICML.
[31] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Daniel Jurafsky,et al. Learning to Decode for Future Success , 2017, ArXiv.
[33] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[34] Yoshua Bengio,et al. Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.
[35] Jiebo Luo,et al. Image Captioning with Semantic Attention , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[37] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[38] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[39] L. Bottou. Curiously Fast Convergence of some Stochastic Gradient Descent Algorithms , 2009 .
[40] Inchul Song,et al. RNNDROP: A novel dropout for RNNS in ASR , 2015, 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU).
[41] Chris Dyer,et al. On the State of the Art of Evaluation in Neural Language Models , 2017, ICLR.
[42] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[43] Nicolas Usunier,et al. Improving Neural Language Models with a Continuous Cache , 2016, ICLR.
[44] Quoc V. Le,et al. Sequence to Sequence Learning with Neural Networks , 2014, NIPS.
[45] Roland Memisevic,et al. Regularizing RNNs by Stabilizing Activations , 2015, ICLR.
[46] Yoshua Bengio,et al. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.
[47] Richard Socher,et al. Knowing When to Look: Adaptive Attention via a Visual Sentinel for Image Captioning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[48] Sepp Hochreiter,et al. Untersuchungen zu dynamischen neuronalen Netzen , 1991 .
[49] Yoshua Bengio,et al. Task Loss Estimation for Sequence Prediction , 2015, ArXiv.
[50] Hugo Larochelle,et al. Neural Autoregressive Distribution Estimation , 2016, J. Mach. Learn. Res..
[51] Jason Weston,et al. Reading Wikipedia to Answer Open-Domain Questions , 2017, ACL.
[52] Tao Mei,et al. Boosting Image Captioning with Attributes , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[53] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[54] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[55] Max Welling,et al. Markov Chain Monte Carlo and Variational Inference: Bridging the Gap , 2014, ICML.
[56] Tom Schaul,et al. The Predictron: End-To-End Learning and Planning , 2016, ICML.
[57] Yoshua Bengio,et al. Plan, Attend, Generate: Planning for Sequence-to-Sequence Models , 2017, NIPS.
[58] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[59] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[60] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[61] Christoph Meinel,et al. Image Captioning with Deep Bidirectional LSTMs , 2016, ACM Multimedia.
[62] Rongrong Ji,et al. Asynchronous Bidirectional Decoding for Neural Machine Translation , 2018, AAAI.
[63] Kuldip K. Paliwal,et al. Bidirectional recurrent neural networks , 1997, IEEE Trans. Signal Process..
[64] Lalit R. Bahl,et al. A Maximum Likelihood Approach to Continuous Speech Recognition , 1983, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[65] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.