On the Spectral Bias of Neural Networks
暂无分享,去创建一个
Yoshua Bengio | Min Lin | Aaron C. Courville | Fred A. Hamprecht | Devansh Arpit | Nasim Rahaman | Aristide Baratin | Felix Dräxler | Yoshua Bengio | Min Lin | Devansh Arpit | A. Baratin | F. Hamprecht | Nasim Rahaman | Felix Dräxler
[1] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..
[2] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[3] Andrew R. Barron,et al. Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.
[4] Allan Pinkus,et al. Multilayer Feedforward Networks with a Non-Polynomial Activation Function Can Approximate Any Function , 1991, Neural Networks.
[5] G. Lugosi,et al. Consistency of the k-Nearest Neighbor Rule , 1996 .
[6] E. Candès. Harmonic Analysis of Neural Networks , 1999 .
[7] Barbara Hammer,et al. A Note on the Universal Approximation Capability of Support Vector Machines , 2003, Neural Processing Letters.
[8] P. Bartlett,et al. Local Rademacher complexities , 2005, math/0508275.
[9] Joachim M. Buhmann,et al. Model Selection in Kernel Methods Based on a Spectral Analysis of Label Information , 2006, DAGM-Symposium.
[10] Carl E. Rasmussen,et al. Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.
[11] G. Fasshauer. Positive definite kernels: past, present and future , 2011 .
[12] E FasshauerG. Positive definite kernels: past, present and future , 2011 .
[13] Razvan Pascanu,et al. On the Number of Linear Regions of Deep Neural Networks , 2014, NIPS.
[14] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[15] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[16] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[17] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[18] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[19] S. Robins,et al. Fourier transforms of polytopes, solid angle sums, and discrete volume , 2016, 1602.08593.
[20] E. L. Kolsbjerg,et al. An automated nudged elastic band method. , 2016, The Journal of chemical physics.
[21] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Matus Telgarsky,et al. Benefits of Depth in Neural Networks , 2016, COLT.
[23] Surya Ganguli,et al. Exponential expressivity in deep neural networks through transient chaos , 2016, NIPS.
[24] Ohad Shamir,et al. The Power of Depth for Feedforward Neural Networks , 2015, COLT.
[25] Surya Ganguli,et al. On the Expressive Power of Deep Neural Networks , 2016, ICML.
[26] Gintare Karolina Dziugaite,et al. Computing Nonvacuous Generalization Bounds for Deep (Stochastic) Neural Networks with Many More Parameters than Training Data , 2017, UAI.
[27] Liwei Wang,et al. The Expressive Power of Neural Networks: A View from the Width , 2017, NIPS.
[28] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[29] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[30] Jorge Nocedal,et al. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.
[31] Raman Arora,et al. Understanding Deep Neural Networks with Rectified Linear Units , 2016, Electron. Colloquium Comput. Complex..
[32] V. Serov. Fourier Series, Fourier Transform and Their Applications to Mathematical Physics , 2017 .
[33] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[34] Mikhail Belkin,et al. Diving into the shallows: a computational perspective on large-scale shallow learning , 2017, NIPS.
[35] David M. Blei,et al. Stochastic Gradient Descent as Approximate Bayesian Inference , 2017, J. Mach. Learn. Res..
[36] T. Poggio,et al. Theory of Deep Learning III : the non-overfitting puzzle , 2018 .
[37] Nathan Srebro,et al. The Implicit Bias of Gradient Descent on Separable Data , 2017, J. Mach. Learn. Res..
[38] Fred A. Hamprecht,et al. Essentially No Barriers in Neural Network Energy Landscape , 2018, ICML.
[39] Hongyi Zhang,et al. mixup: Beyond Empirical Risk Minimization , 2017, ICLR.
[40] Yoshua Bengio,et al. A Walk with SGD , 2018, ArXiv.
[41] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[42] Stefano Soatto,et al. Stochastic Gradient Descent Performs Variational Inference, Converges to Limit Cycles for Deep Networks , 2017, 2018 Information Theory and Applications Workshop (ITA).
[43] Quoc V. Le,et al. A Bayesian Perspective on Generalization and Stochastic Gradient Descent , 2017, ICLR.
[44] Liwen Zhang,et al. Tropical Geometry of Deep Neural Networks , 2018, ICML.
[45] Zhi-Qin John Xu,et al. Understanding training and generalization in deep learning by Fourier analysis , 2018, ArXiv.
[46] Jascha Sohl-Dickstein,et al. Sensitivity and Generalization in Neural Networks: an Empirical Study , 2018, ICLR.
[47] Michael Unser,et al. A representer theorem for deep neural networks , 2018, J. Mach. Learn. Res..
[48] Zhi-Qin John Xu,et al. Training behavior of deep neural network in frequency domain , 2018, ICONIP.
[49] M. Spivak. Calculus On Manifolds: A Modern Approach To Classical Theorems Of Advanced Calculus , 2019 .
[50] Andrew M. Saxe,et al. High-dimensional dynamics of generalization error in neural networks , 2017, Neural Networks.