Diverse Neural Network Learns True Target Functions
暂无分享,去创建一个
Le Song | Yingyu Liang | Bo Xie | Le Song | Bo Xie | Yingyu Liang
[1] Ronald L. Rivest,et al. Training a 3-node neural network is NP-complete , 1988, COLT '88.
[2] Andrew R. Barron,et al. Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.
[3] Claus Müller. Analysis of Spherical Symmetries in Euclidean Spaces , 1997 .
[4] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[5] Mikio L. Braun,et al. Accurate Error Bounds for the Eigenvalues of the Kernel Matrix , 2006, J. Mach. Learn. Res..
[6] AI Koan,et al. Weighted Sums of Random Kitchen Sinks: Replacing minimization with randomization in learning , 2008, NIPS.
[7] Lawrence K. Saul,et al. Kernel Methods for Deep Learning , 2009, NIPS.
[8] Liva Ralaivola,et al. Empirical Bernstein Inequalities for U-Statistics , 2010, NIPS.
[9] Roman Vershynin,et al. Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.
[10] J. Tropp. User-Friendly Tools for Random Matrices: An Introduction , 2012 .
[11] Surya Ganguli,et al. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization , 2014, NIPS.
[12] D. Bilyk,et al. One Bit Sensing, Discrepancy, and Stolarsky Principle , 2015 .
[13] Pengtao Xie,et al. On the Generalization Error Bounds of Neural Networks under Diversity-Inducing Mutual Angular Regularization , 2015, ArXiv.
[14] Furong Huang,et al. Escaping From Saddle Points - Online Stochastic Gradient for Tensor Decomposition , 2015, COLT.
[15] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[16] Kenji Kawaguchi,et al. Deep Learning without Poor Local Minima , 2016, NIPS.
[17] Daniel Soudry,et al. No bad local minima: Data independent training error guarantees for multilayer neural networks , 2016, ArXiv.
[18] Michael I. Jordan,et al. Gradient Descent Only Converges to Minimizers , 2016, COLT.
[19] Tengyu Ma,et al. Matrix Completion has No Spurious Local Minimum , 2016, NIPS.
[20] Lior Wolf,et al. The Multiverse Loss for Robust Transfer Learning , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] John J. Hopfield,et al. Dense Associative Memory for Pattern Recognition , 2016, NIPS.
[22] Anima Anandkumar,et al. Beating the Perils of Non-Convexity: Guaranteed Training of Neural Networks using Tensor Methods , 2017 .
[23] Ohad Shamir,et al. Distribution-Specific Hardness of Learning Neural Networks , 2016, J. Mach. Learn. Res..