暂无分享,去创建一个
[1] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[2] Surya Ganguli,et al. On the saddle point problem for non-convex optimization , 2014, ArXiv.
[3] Yann LeCun,et al. Handwritten zip code recognition with multilayer networks , 1990, [1990] Proceedings. 10th International Conference on Pattern Recognition.
[4] Oriol Vinyals,et al. Qualitatively characterizing neural network optimization problems , 2014, ICLR.
[5] Kenji Kawaguchi,et al. Deep Learning without Poor Local Minima , 2016, NIPS.
[6] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[7] Vivienne Sze,et al. Efficient Processing of Deep Neural Networks: A Tutorial and Survey , 2017, Proceedings of the IEEE.
[8] Abraham J. Wyner,et al. Modern Neural Networks Generalize on Small Data Sets , 2018, NeurIPS.
[9] Leslie Pack Kaelbling,et al. Generalization in Deep Learning , 2017, ArXiv.
[10] Yoram Singer,et al. Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.
[11] Lei Wu,et al. Towards Understanding Generalization of Deep Learning: Perspective of Loss Landscapes , 2017, ArXiv.
[12] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[13] Yuanzhi Li,et al. Learning and Generalization in Overparameterized Neural Networks, Going Beyond Two Layers , 2018, NeurIPS.
[14] Brendan J. Frey,et al. Adaptive dropout for training deep neural networks , 2013, NIPS.
[15] Aleksander Madry,et al. How Does Batch Normalization Help Optimization? (No, It Is Not About Internal Covariate Shift) , 2018, NeurIPS.
[16] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[17] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[18] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[19] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[20] Surya Ganguli,et al. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization , 2014, NIPS.
[21] Surya Ganguli,et al. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.
[22] Jacques Bughin,et al. A future that works: automation, employment, and productivity , 2017 .