暂无分享,去创建一个
[1] J. Zico Kolter,et al. Uniform convergence may be unable to explain generalization in deep learning , 2019, NeurIPS.
[2] Gintare Karolina Dziugaite,et al. Computing Nonvacuous Generalization Bounds for Deep (Stochastic) Neural Networks with Many More Parameters than Training Data , 2017, UAI.
[3] Xin Dong,et al. Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon , 2017, NIPS.
[4] Frederik Kunstner,et al. Limitations of the Empirical Fisher Approximation , 2019, NeurIPS.
[5] Shiliang Sun,et al. PAC-bayes bounds with data dependent priors , 2012, J. Mach. Learn. Res..
[6] Leslie Pack Kaelbling,et al. Generalization in Deep Learning , 2017, ArXiv.
[7] Tomaso A. Poggio,et al. Fisher-Rao Metric, Geometry, and Complexity of Neural Networks , 2017, AISTATS.
[8] Ilya Sutskever,et al. Estimating the Hessian by Back-propagating Curvature , 2012, ICML.
[9] Gintare Karolina Dziugaite,et al. Entropy-SGD optimizes the prior of a PAC-Bayes bound: Data-dependent PAC-Bayes priors via differential privacy , 2017, NeurIPS.
[10] Sebastian Nowozin,et al. Deterministic Variational Inference for Robust Bayesian Neural Networks , 2018, ICLR.
[11] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[12] Jiaxiang Wu,et al. Collaborative Channel Pruning for Deep Networks , 2019, ICML.
[13] Nicolas Le Roux,et al. Information matrices and generalization , 2019, ArXiv.
[14] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[15] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[16] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[17] Sanja Fidler,et al. EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis , 2019, ICML.
[18] Stefano Soatto,et al. The Information Complexity of Learning Tasks, their Structure and their Distance , 2019, Information and Inference: A Journal of the IMA.
[19] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[20] Hossein Mobahi,et al. Predicting the Generalization Gap in Deep Networks with Margin Distributions , 2018, ICLR.
[21] Trac D. Tran,et al. A Scale Invariant Flatness Measure for Deep Network Minima , 2019, ArXiv.
[22] O. Catoni. A PAC-Bayesian approach to adaptive classification , 2004 .
[23] John Shawe-Taylor,et al. Tighter PAC-Bayes Bounds , 2006, NIPS.
[24] Jascha Sohl-Dickstein,et al. Sensitivity and Generalization in Neural Networks: an Empirical Study , 2018, ICLR.
[25] Ohad Shamir,et al. Size-Independent Sample Complexity of Neural Networks , 2017, COLT.
[26] David A. McAllester. Some PAC-Bayesian Theorems , 1998, COLT' 98.
[27] David A. McAllester,et al. A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks , 2017, ICLR.
[28] Cynthia Dwork,et al. Differential Privacy , 2006, ICALP.
[29] Radford M. Neal. Pattern Recognition and Machine Learning , 2007, Technometrics.
[30] Razvan Pascanu,et al. Sharp Minima Can Generalize For Deep Nets , 2017, ICML.
[31] Andreas Loukas,et al. Some limitations of norm based generalization bounds in deep neural networks , 2019, ArXiv.
[32] Ryan P. Adams,et al. Non-vacuous Generalization Bounds at the ImageNet Scale: a PAC-Bayesian Compression Approach , 2018, ICLR.
[33] Naftali Tishby,et al. The information bottleneck method , 2000, ArXiv.
[34] Stefano Soatto,et al. Emergence of Invariance and Disentanglement in Deep Representations , 2017, 2018 Information Theory and Applications Workshop (ITA).
[35] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[36] Roger B. Grosse,et al. Optimizing Neural Networks with Kronecker-factored Approximate Curvature , 2015, ICML.