暂无分享,去创建一个
Yann LeCun | Zhiyuan Li | Nathan Srebro | Behnam Neyshabur | Srinadh Bhojanapalli | Yann LeCun | Nathan Srebro | Srinadh Bhojanapalli | Behnam Neyshabur | Zhiyuan Li | N. Srebro
[1] M. Talagrand,et al. Probability in Banach Spaces: Isoperimetry and Processes , 1991 .
[2] C. Lee Giles,et al. What Size Neural Network Gives Optimal Generalization? Convergence Properties of Backpropagation , 1998 .
[3] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[4] Nicolas Le Roux,et al. Convex Neural Networks , 2005, NIPS.
[5] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[6] Ameet Talwalkar,et al. Foundations of Machine Learning , 2012, Adaptive computation and machine learning.
[7] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[8] Andreas Krause,et al. Advances in Neural Information Processing Systems (NIPS) , 2014 .
[9] Ryota Tomioka,et al. Norm-Based Capacity Control in Neural Networks , 2015, COLT.
[10] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[11] Ruslan Salakhutdinov,et al. Path-SGD: Path-Normalized Optimization in Deep Neural Networks , 2015, NIPS.
[12] Andreas Maurer,et al. A Vector-Contraction Inequality for Rademacher Complexities , 2016, ALT.
[13] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Gintare Karolina Dziugaite,et al. Computing Nonvacuous Generalization Bounds for Deep (Stochastic) Neural Networks with Many More Parameters than Training Data , 2017, UAI.
[15] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[16] Jorge Nocedal,et al. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.
[17] Abbas Mehrabian,et al. Nearly-tight VC-dimension bounds for piecewise linear neural networks , 2017, COLT.
[18] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[19] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[20] Francis R. Bach,et al. Breaking the Curse of Dimensionality with Convex Neural Networks , 2014, J. Mach. Learn. Res..
[21] Nathan Srebro,et al. The Implicit Bias of Gradient Descent on Separable Data , 2017, J. Mach. Learn. Res..
[22] Ohad Shamir,et al. Size-Independent Sample Complexity of Neural Networks , 2017, COLT.
[23] Jaehoon Lee,et al. Deep Neural Networks as Gaussian Processes , 2017, ICLR.
[24] Nathan Srebro,et al. Implicit Regularization in Matrix Factorization , 2017, 2018 Information Theory and Applications Workshop (ITA).
[25] David A. McAllester,et al. A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks , 2017, ICLR.
[26] Yi Zhang,et al. Stronger generalization bounds for deep nets via a compression approach , 2018, ICML.
[27] Jascha Sohl-Dickstein,et al. Sensitivity and Generalization in Neural Networks: an Empirical Study , 2018, ICLR.
[28] Peter L. Bartlett,et al. Nearly-tight VC-dimension and Pseudodimension Bounds for Piecewise Linear Neural Networks , 2017, J. Mach. Learn. Res..
[29] J. Zico Kolter,et al. Generalization in Deep Networks: The Role of Distance from Initialization , 2019, ArXiv.
[30] Tomaso A. Poggio,et al. Fisher-Rao Metric, Geometry, and Complexity of Neural Networks , 2017, AISTATS.