Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks
暂无分享,去创建一个
[1] Boris Polyak,et al. Acceleration of stochastic approximation by averaging , 1992 .
[2] Shun-ichi Amari,et al. Neural Learning in Structured Parameter Spaces - Natural Riemannian Gradient , 1996, NIPS.
[3] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[4] Adi Shraibman,et al. Rank, Trace-Norm and Max-Norm , 2005, COLT.
[5] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[6] James Martens,et al. Deep learning via Hessian-free optimization , 2010, ICML.
[7] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[8] Razvan Pascanu,et al. Theano: new features and speed improvements , 2012, ArXiv.
[9] Tapani Raiko,et al. Deep Learning Made Easier by Linear Transformations in Perceptrons , 2012, AISTATS.
[10] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[11] Yoshua Bengio,et al. Maxout Networks , 2013, ICML.
[12] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[13] Qiang Chen,et al. Network In Network , 2013, ICLR.
[14] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[15] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[16] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[17] Ruslan Salakhutdinov,et al. Scaling up Natural Gradient by Sparsely Factorizing the Inverse Fisher Matrix , 2015, ICML.
[18] Max Welling,et al. Markov Chain Monte Carlo and Variational Inference: Bridging the Gap , 2014, ICML.
[19] Razvan Pascanu,et al. Natural Neural Networks , 2015, NIPS.
[20] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[21] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[22] Roger B. Grosse,et al. Optimizing Neural Networks with Kronecker-factored Approximate Curvature , 2015, ICML.
[23] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[24] Shiliang Zhang,et al. Rectified linear neural networks with tied-scalar regularization for LVCSR , 2015, INTERSPEECH.
[25] Marc G. Bellemare,et al. The Arcade Learning Environment: An Evaluation Platform for General Agents , 2012, J. Artif. Intell. Res..
[26] Zhuowen Tu,et al. Deeply-Supervised Nets , 2014, AISTATS.
[27] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[29] Jiri Matas,et al. All you need is a good init , 2015, ICLR.
[30] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[31] Trevor Darrell,et al. Data-dependent Initializations of Convolutional Neural Networks , 2015, ICLR.
[32] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Max Welling,et al. Improved Variational Inference with Inverse Autoregressive Flow , 2016, NIPS 2016.
[34] Aaron C. Courville,et al. Recurrent Batch Normalization , 2016, ICLR.