Are deep ResNets provably better than linear predictors?
暂无分享,去创建一个
[1] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[2] Ohad Shamir,et al. Size-Independent Sample Complexity of Neural Networks , 2017, COLT.
[3] Matthias Hein,et al. The Loss Surface of Deep and Wide Neural Networks , 2017, ICML.
[4] Suvrit Sra,et al. Global optimality conditions for deep neural networks , 2017, ICLR.
[5] Stefanie Jegelka,et al. ResNet with one-neuron hidden layers is a Universal Approximator , 2018, NeurIPS.
[6] Hao Li,et al. Visualizing the Loss Landscape of Neural Nets , 2017, NeurIPS.
[7] Ohad Shamir,et al. Are ResNets Provably Better than Linear Predictors? , 2018, NeurIPS.
[8] Tengyu Ma,et al. Identity Matters in Deep Learning , 2016, ICLR.
[9] Thomas Laurent,et al. The Multilinear Structure of ReLU Networks , 2017, ICML.
[10] Le Song,et al. Diverse Neural Network Learns True Target Functions , 2016, AISTATS.
[11] Daniel Soudry,et al. No bad local minima: Data independent training error guarantees for multilayer neural networks , 2016, ArXiv.
[12] R. Srikant,et al. Understanding the Loss Surface of Neural Networks for Binary Classification , 2018, ICML.
[13] Philip M. Long,et al. Representing smooth functions as compositions of near-identity functions with implications for deep network optimization , 2018, ArXiv.
[14] Jason D. Lee,et al. No Spurious Local Minima in a Two Hidden Unit ReLU Network , 2018, ICLR.
[15] R. Srikant,et al. Adding One Neuron Can Eliminate All Bad Local Minima , 2018, NeurIPS.
[16] Yoshua Bengio,et al. Depth with Nonlinearity Creates No Bad Local Minima in ResNets , 2019, Neural Networks.
[17] Suvrit Sra,et al. Efficiently testing local optimality and escaping saddles for ReLU networks , 2018, ICLR.
[18] Ryota Tomioka,et al. Norm-Based Capacity Control in Neural Networks , 2015, COLT.
[19] Suvrit Sra,et al. Small nonlinearities in activation functions create bad local minima in neural networks , 2018, ICLR.
[20] X H Yu,et al. On the local minima free condition of backpropagation learning , 1995, IEEE Trans. Neural Networks.
[21] Razvan Pascanu,et al. Local minima in training of neural networks , 2016, 1611.06310.
[22] Thomas Laurent,et al. Deep Linear Networks with Arbitrary Loss: All Local Minima Are Global , 2017, ICML.
[23] Matthias Hein,et al. On the loss landscape of a class of deep neural networks with no bad local valleys , 2018, ICLR.
[24] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[25] Ohad Shamir,et al. Spurious Local Minima are Common in Two-Layer ReLU Neural Networks , 2017, ICML.
[26] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Yi Zhou,et al. Critical Points of Linear Neural Networks: Analytical Forms and Landscape Properties , 2017, ICLR.
[28] Kenji Kawaguchi,et al. Deep Learning without Poor Local Minima , 2016, NIPS.
[29] Yi Zhou,et al. Critical Points of Neural Networks: Analytical Forms and Landscape Properties , 2017, ArXiv.
[30] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[31] Tengyu Ma,et al. Fixup Initialization: Residual Learning Without Normalization , 2019, ICLR.
[32] Kurt Hornik,et al. Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.