暂无分享,去创建一个
Yi Zhang | Sanjeev Arora | Behnam Neyshabur | Rong Ge | Sanjeev Arora | Rong Ge | Behnam Neyshabur | Yi Zhang
[1] R. Dudley. Universal Donsker Classes and Metric Entropy , 1987 .
[2] Geoffrey E. Hinton,et al. Keeping the neural networks simple by minimizing the description length of the weights , 1993, COLT '93.
[3] Jürgen Schmidhuber,et al. Flat Minima , 1997, Neural Computation.
[4] Bernhard Schölkopf,et al. The connection between regularization operators and support vector kernels , 1998, Neural Networks.
[5] Dana Ron,et al. Algorithmic Stability and Sanity-Check Bounds for Leave-One-Out Cross-Validation , 1997, Neural Computation.
[6] David A. McAllester. PAC-Bayesian model averaging , 1999, COLT '99.
[7] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[8] Tomaso A. Poggio,et al. Regularization Networks and Support Vector Machines , 2000, Adv. Comput. Math..
[9] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[10] John Langford,et al. (Not) Bounding the True Error , 2001, NIPS.
[11] Stephen P. Boyd,et al. A rank minimization heuristic with application to minimum order system approximation , 2001, Proceedings of the 2001 American Control Conference. (Cat. No.01CH37148).
[12] Manfred K. Warmuth,et al. Relating Data Compression and Learnability , 2003 .
[13] David A. McAllester. Some PAC-Bayesian Theorems , 1998, COLT' 98.
[14] Tommi S. Jaakkola,et al. Maximum-Margin Matrix Factorization , 2004, NIPS.
[15] Avrim Blum,et al. Random Projection, Margins, Kernels, and Feature-Selection , 2005, SLSFS.
[16] Ohad Shamir,et al. Learnability, Stability and Uniform Convergence , 2010, J. Mach. Learn. Res..
[17] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[18] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[19] Ryota Tomioka,et al. Norm-Based Capacity Control in Neural Networks , 2015, COLT.
[20] J. Ramon,et al. Hoeffding's inequality for sums of weakly dependent random variables , 2015, 1507.06871.
[21] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[22] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[23] Ruslan Salakhutdinov,et al. Path-SGD: Path-Normalized Optimization in Deep Neural Networks , 2015, NIPS.
[24] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[25] Yoram Singer,et al. Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.
[26] Leslie Pack Kaelbling,et al. Generalization in Deep Learning , 2017, ArXiv.
[27] Gintare Karolina Dziugaite,et al. Computing Nonvacuous Generalization Bounds for Deep (Stochastic) Neural Networks with Many More Parameters than Training Data , 2017, UAI.
[28] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[29] Razvan Pascanu,et al. Sharp Minima Can Generalize For Deep Nets , 2017, ICML.
[30] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[31] Jorge Nocedal,et al. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.
[32] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[33] Stefano Soatto,et al. Entropy-SGD: biasing gradient descent into wide valleys , 2016, ICLR.
[34] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[35] Ohad Shamir,et al. Size-Independent Sample Complexity of Neural Networks , 2017, COLT.
[36] Matthew Botvinick,et al. On the importance of single directions for generalization , 2018, ICLR.
[37] David A. McAllester,et al. A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks , 2017, ICLR.
[38] Tao Zhang,et al. Model Compression and Acceleration for Deep Neural Networks: The Principles, Progress, and Challenges , 2018, IEEE Signal Processing Magazine.
[39] Tomaso A. Poggio,et al. Fisher-Rao Metric, Geometry, and Complexity of Neural Networks , 2017, AISTATS.