A Simple Baseline for Bayesian Uncertainty in Deep Learning
暂无分享,去创建一个
Andrew Gordon Wilson | Dmitry P. Vetrov | Pavel Izmailov | Wesley J. Maddox | Timur Garipov | Wesley Maddox | D. Vetrov | Pavel Izmailov | T. Garipov | A. Wilson
[1] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[2] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[3] David J. C. MacKay,et al. Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.
[4] J. Berger. Statistical Decision Theory and Bayesian Analysis , 1988 .
[5] Alex Kendall,et al. What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.
[6] Ethan Dyer,et al. Gradient Descent Happens in a Tiny Subspace , 2018, ArXiv.
[7] Ulrich K. Müller. RISK OF BAYESIAN INFERENCE IN MISSPECIFIED MODELS, AND THE SANDWICH COVARIANCE MATRIX , 2013 .
[8] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[9] Jishnu Mukhoti,et al. Evaluating Bayesian Deep Learning Methods for Semantic Segmentation , 2018, ArXiv.
[10] Stefano Soatto,et al. Stochastic Gradient Descent Performs Variational Inference, Converges to Limit Cycles for Deep Networks , 2017, 2018 Information Theory and Applications Workshop (ITA).
[11] Stefan Winkler,et al. The Unusual Effectiveness of Averaging in GAN Training , 2018, ICLR.
[12] David J. C. MacKay,et al. A Practical Bayesian Framework for Backpropagation Networks , 1992, Neural Computation.
[13] Andrew Gordon Wilson,et al. Cyclical Stochastic Gradient MCMC for Bayesian Deep Learning , 2019, ICLR.
[14] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[15] Hao Li,et al. Visualizing the Loss Landscape of Neural Nets , 2017, NeurIPS.
[16] Andrew Gordon Wilson,et al. Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs , 2018, NeurIPS.
[17] Frank Hutter,et al. Decoupled Weight Decay Regularization , 2017, ICLR.
[18] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[19] David M. Blei,et al. Stochastic Gradient Descent as Approximate Bayesian Inference , 2017, J. Mach. Learn. Res..
[20] A. G. Wilson,et al. Fast Uncertainty Estimates and Bayesian Model Averaging of DNNs , 2018 .
[21] David J. C. MacKay,et al. Bayesian Interpolation , 1992, Neural Computation.
[22] Didrik Nielsen,et al. Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam , 2018, ICML.
[23] Razvan Pascanu,et al. Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.
[24] Alex Kendall,et al. Concrete Dropout , 2017, NIPS.
[25] Max Welling,et al. Multiplicative Normalizing Flows for Variational Bayesian Neural Networks , 2017, ICML.
[26] Andrew Gordon Wilson,et al. Averaging Weights Leads to Wider Optima and Better Generalization , 2018, UAI.
[27] Guodong Zhang,et al. Noisy Natural Gradient as Variational Inference , 2017, ICML.
[28] Geoffrey E. Hinton,et al. Bayesian Learning for Neural Networks , 1995 .
[29] Julien Cornebise,et al. Weight Uncertainty in Neural Network , 2015, ICML.
[30] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Appendix , 2015, 1506.02157.
[31] Ohad Shamir,et al. Making Gradient Descent Optimal for Strongly Convex Stochastic Optimization , 2011, ICML.
[32] David Barber,et al. Online Structured Laplace Approximations For Overcoming Catastrophic Forgetting , 2018, NeurIPS.
[33] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[34] Boris Polyak,et al. Acceleration of stochastic approximation by averaging , 1992 .
[35] Sanjoy Dasgupta,et al. On-Line Estimation with the Multivariate Gaussian Distribution , 2007, COLT.
[36] Andrew Gordon Wilson,et al. There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average , 2018, ICLR.
[37] Tianqi Chen,et al. Stochastic Gradient Hamiltonian Monte Carlo , 2014, ICML.
[38] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[39] Xin T. Tong,et al. Statistical inference for model parameters in stochastic gradient descent , 2016, The Annals of Statistics.
[40] Andrew Gordon Wilson,et al. GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration , 2018, NeurIPS.
[41] Yann Dauphin,et al. Empirical Analysis of the Hessian of Over-Parametrized Neural Networks , 2017, ICLR.
[42] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[43] Fred A. Hamprecht,et al. Essentially No Barriers in Neural Network Energy Landscape , 2018, ICML.
[44] Daniel Hernández-Lobato,et al. Deep Gaussian Processes for Regression using Approximate Expectation Propagation , 2016, ICML.
[45] Ryan P. Adams,et al. Probabilistic Backpropagation for Scalable Learning of Bayesian Neural Networks , 2015, ICML.
[46] Julien Cornebise,et al. Weight Uncertainty in Neural Networks , 2015, ArXiv.
[47] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[48] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[49] Nathan Halko,et al. Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions , 2009, SIAM Rev..
[50] Jorge Nocedal,et al. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.
[51] Honglak Lee,et al. An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.
[52] Yee Whye Teh,et al. Bayesian Learning via Stochastic Gradient Langevin Dynamics , 2011, ICML.
[53] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[54] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[55] Francis Bach,et al. Constant Step Size Stochastic Gradient Descent for Probabilistic Modeling , 2018, UAI.
[56] Yann Ollivier,et al. The Description Length of Deep Learning models , 2018, NeurIPS.
[57] David Barber,et al. A Scalable Laplace Approximation for Neural Networks , 2018, ICLR.
[58] Sebastian Nowozin,et al. Deterministic Variational Inference for Robust Bayesian Neural Networks , 2018, ICLR.
[59] Andrew Gordon Wilson,et al. Subspace Inference for Bayesian Deep Learning , 2019, UAI.
[60] Andrew Gordon Wilson,et al. Bayesian GAN , 2017, NIPS.
[61] Stefano Ermon,et al. Accurate Uncertainties for Deep Learning Using Calibrated Regression , 2018, ICML.
[62] Christopher De Sa,et al. SWALP : Stochastic Weight Averaging in Low-Precision Training , 2019, ICML.
[63] Rich Caruana,et al. Predicting good probabilities with supervised learning , 2005, ICML.
[64] A. G. Wilson,et al. Improving Stability in Deep Reinforcement Learning with Weight Averaging , 2018 .
[65] Sebastian Nowozin,et al. Fixing Variational Bayes: Deterministic Variational Inference for Bayesian Neural Networks , 2018, ArXiv.
[66] Milos Hauskrecht,et al. Obtaining Well Calibrated Probabilities Using Bayesian Binning , 2015, AAAI.