Experimental Comparison of Stochastic Optimizers in Deep Learning
暂无分享,去创建一个
Emmanuel Okewu | Philip Adewole | Oladipupo Sennaike | Oladipupo A. Sennaike | E. Okewu | P. Adewole
[1] Chao Wang,et al. On global convergence of gradient descent algorithms for generalized phase retrieval problem , 2018, J. Comput. Appl. Math..
[2] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[3] Geoffrey E. Hinton,et al. Training Recurrent Neural Networks , 2013 .
[4] Jürgen Schmidhuber,et al. Deep learning in neural networks: An overview , 2014, Neural Networks.
[5] Hiroaki Hayashi,et al. Improving Stochastic Gradient Descent with Feedback , 2016, ArXiv.
[6] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[7] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[8] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[9] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[10] Alex Graves,et al. Generating Sequences With Recurrent Neural Networks , 2013, ArXiv.
[11] Alexandros Iosifidis,et al. Improving Efficiency in Convolutional Neural Network with Multilinear Filters , 2018, Neural Networks.
[12] Yoram Singer,et al. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..
[13] Andrea Montanari,et al. A mean field view of the landscape of two-layer neural networks , 2018, Proceedings of the National Academy of Sciences.
[14] Lei Deng,et al. Training deep neural networks with discrete state transition , 2018, Neurocomputing.
[15] Donghwan Kim,et al. Optimized first-order methods for smooth convex minimization , 2014, Mathematical Programming.