暂无分享,去创建一个
Yoshua Bengio | Devansh Arpit | Dendi Suhubdy | Konrad Zolna | Yoshua Bengio | Devansh Arpit | Konrad Zolna | Dendi Suhubdy
[1] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[2] Tolga Tasdizen,et al. Regularization With Stochastic Transformations and Perturbations for Deep Semi-Supervised Learning , 2016, NIPS.
[3] David M. Blei,et al. Stochastic Gradient Descent as Approximate Bayesian Inference , 2017, J. Mach. Learn. Res..
[4] Philipp Koehn,et al. Moses: Open Source Toolkit for Statistical Machine Translation , 2007, ACL.
[5] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[6] Yoshua Bengio,et al. Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations , 2016, ICLR.
[7] Tapani Raiko,et al. Semi-supervised Learning with Ladder Networks , 2015, NIPS.
[8] Hakan Inan,et al. Tying Word Vectors and Word Classifiers: A Loss Framework for Language Modeling , 2016, ICLR.
[9] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[10] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[11] Timo Aila,et al. Temporal Ensembling for Semi-Supervised Learning , 2016, ICLR.
[12] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[13] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[14] Richard Socher,et al. Revisiting Activation Regularization for Language RNNs , 2017, ArXiv.
[15] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[17] Richard Socher,et al. Pointer Sentinel Mixture Models , 2016, ICLR.
[18] Jürgen Schmidhuber,et al. Recurrent Highway Networks , 2016, ICML.
[19] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[20] Yaoliang Yu,et al. Dropout with Expectation-linear Regularization , 2016, ICLR.
[21] Boris Polyak,et al. Acceleration of stochastic approximation by averaging , 1992 .
[22] Chris Dyer,et al. On the State of the Art of Evaluation in Neural Language Models , 2017, ICLR.
[23] Samy Bengio,et al. Show and tell: A neural image caption generator , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[24] Lior Wolf,et al. Using the Output Embedding to Improve Language Models , 2016, EACL.
[25] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[26] Christopher Joseph Pal,et al. Twin Networks: Matching the Future for Sequence Generation , 2017, ICLR.
[27] Ying Zhang,et al. Batch normalized recurrent neural networks , 2015, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[28] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[29] Aaron C. Courville,et al. Recurrent Batch Normalization , 2016, ICLR.