暂无分享,去创建一个
[1] Kevin Smith,et al. Bayesian Uncertainty Estimation for Batch Normalized Deep Networks , 2018, ICML.
[2] P. A. Ligomenides,et al. Uncertainty in neural networks , 1993, 1993 (2nd) International Symposium on Uncertainty Modeling and Analysis.
[3] Yarin Gal,et al. Uncertainty in Deep Learning , 2016 .
[4] Jasper Snoek,et al. Hyperparameter Ensembles for Robustness and Uncertainty Quantification , 2020, NeurIPS.
[5] Bo Zhang,et al. Function Space Particle Optimization for Bayesian Neural Networks , 2019, ICLR.
[6] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[7] John Schulman,et al. Concrete Problems in AI Safety , 2016, ArXiv.
[8] Matthias Hein,et al. Towards neural networks that provably know when they don't know , 2020, ICLR.
[9] Andrew Gordon Wilson,et al. Subspace Inference for Bayesian Deep Learning , 2019, UAI.
[10] Yee Whye Teh,et al. Simple and Scalable Epistemic Uncertainty Estimation Using a Single Deep Deterministic Neural Network , 2020, ICML.
[11] Andrew Gordon Wilson,et al. A Simple Baseline for Bayesian Uncertainty in Deep Learning , 2019, NeurIPS.
[12] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[13] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[14] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[15] Justin Bayer,et al. Bayesian Learning of Neural Network Architectures , 2019, AISTATS.
[16] Valen E. Johnson,et al. High-Dimensional Bayesian Classifiers Using Non-Local Priors , 2013, Statistical Models for Data Analysis.
[17] Marc Peter Deisenroth,et al. Distributed Gaussian Processes , 2015, ICML.
[18] Andrew Gordon Wilson,et al. The Case for Bayesian Deep Learning , 2020, ArXiv.
[19] Kilian Q. Weinberger,et al. Deep Networks with Stochastic Depth , 2016, ECCV.
[20] Jason Yosinski,et al. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Padhraic Smyth,et al. Dropout as a Structured Shrinkage Prior , 2018, ICML.
[22] V. Johnson,et al. Bayesian Model Selection in High-Dimensional Settings , 2012, Journal of the American Statistical Association.
[23] Richard E. Turner,et al. Pathologies of Factorised Gaussian and MC Dropout Posteriors in Bayesian Neural Networks , 2019, ArXiv.
[24] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[25] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[26] Jeremy Nixon,et al. Measuring Calibration in Deep Learning , 2019, CVPR Workshops.
[27] José Miguel Hernández-Lobato,et al. Variational Implicit Processes , 2018, ICML.
[28] Dustin Tran,et al. Reliable Uncertainty Estimates in Deep Neural Networks using Noise Contrastive Priors , 2018, ArXiv.
[29] Geoffrey E. Hinton,et al. Keeping the neural networks simple by minimizing the description length of the weights , 1993, COLT '93.
[30] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[31] Guodong Zhang,et al. Functional Variational Bayesian Neural Networks , 2019, ICLR.
[32] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[33] Geoffrey E. Hinton,et al. Bayesian Learning for Neural Networks , 1995 .
[34] Paul H. Kupiec,et al. Techniques for Verifying the Accuracy of Risk Measurement Models , 1995 .
[35] Sebastian Nowozin,et al. How Good is the Bayes Posterior in Deep Neural Networks Really? , 2020, ICML.
[36] A. Raftery,et al. Strictly Proper Scoring Rules, Prediction, and Estimation , 2007 .
[37] Federico Tombari,et al. Sampling-Free Epistemic Uncertainty Estimation Using Approximated Variance Propagation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[38] Ameet Talwalkar,et al. Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization , 2016, J. Mach. Learn. Res..
[39] Alex Lamb,et al. Deep Learning for Classical Japanese Literature , 2018, ArXiv.
[40] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[41] Kilian Q. Weinberger,et al. Snapshot Ensembles: Train 1, get M for free , 2017, ICLR.
[42] Soumya Ghosh,et al. Model Selection in Bayesian Neural Networks via Horseshoe Priors , 2017, J. Mach. Learn. Res..
[43] David Barber,et al. A Scalable Laplace Approximation for Neural Networks , 2018, ICLR.
[44] Andrew Gordon Wilson,et al. GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration , 2018, NeurIPS.
[45] Thomas P. Minka,et al. Bayesian model averaging is not model combination , 2002 .
[46] Geoffrey E. Hinton,et al. Analyzing and Improving Representations with the Soft Nearest Neighbor Loss , 2019, ICML.
[47] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[48] Laurence Aitchison,et al. Deep Convolutional Networks as shallow Gaussian Processes , 2018, ICLR.
[49] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[50] Andrew Gordon Wilson,et al. Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs , 2018, NeurIPS.
[51] Neil D. Lawrence. Note Relevance Determination , 2001, WIRN.
[52] Jos'e Miguel Hern'andez-Lobato,et al. Variational Depth Search in ResNets , 2020, ArXiv.
[53] Yee Whye Teh,et al. Hybrid Models with Deep and Invertible Features , 2019, ICML.
[54] Ryan P. Adams,et al. Probabilistic Backpropagation for Scalable Learning of Bayesian Neural Networks , 2015, ICML.
[55] Kaiming He,et al. Accurate, Large Minibatch SGD: Training ImageNet in 1 Hour , 2017, ArXiv.
[56] Didrik Nielsen,et al. Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam , 2018, ICML.
[57] Sebastian Nowozin,et al. Can You Trust Your Model's Uncertainty? Evaluating Predictive Uncertainty Under Dataset Shift , 2019, NeurIPS.
[58] Yee Whye Teh,et al. Do Deep Generative Models Know What They Don't Know? , 2018, ICLR.
[59] Christopher M. Bishop,et al. Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .
[60] Jasper Snoek,et al. Training independent subnetworks for robust prediction , 2020, ArXiv.
[61] Mohammad Emtiyaz Khan,et al. Practical Deep Learning with Bayesian Principles , 2019, NeurIPS.
[62] Aaron Klein,et al. BOHB: Robust and Efficient Hyperparameter Optimization at Scale , 2018, ICML.
[63] Dmitry Vetrov,et al. Pitfalls of In-Domain Uncertainty Estimation and Ensembling in Deep Learning , 2020, ICLR.
[64] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[65] Julien Cornebise,et al. Weight Uncertainty in Neural Network , 2015, ICML.
[66] Yarin Gal,et al. A Systematic Comparison of Bayesian Deep Learning Robustness in Diabetic Retinopathy Tasks , 2019, ArXiv.
[67] Richard E. Turner,et al. Overpruning in Variational Bayesian Neural Networks , 2018, 1801.06230.
[68] Julien Cornebise,et al. Weight Uncertainty in Neural Networks , 2015, ArXiv.
[69] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[70] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[71] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[72] Neil D. Lawrence,et al. Gaussian Processes for Big Data , 2013, UAI.
[73] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[74] Richard E. Turner,et al. Gaussian Process Behaviour in Wide Deep Neural Networks , 2018, ICLR.
[75] Aidan N. Gomez,et al. Benchmarking Bayesian Deep Learning with Diabetic Retinopathy Diagnosis , 2019 .
[76] Jasper Snoek,et al. Efficient and Scalable Bayesian Neural Nets with Rank-1 Factors , 2020, ICML.
[77] David J. C. MacKay,et al. A Practical Bayesian Framework for Backpropagation Networks , 1992, Neural Computation.
[78] David J. C. MacKay,et al. BAYESIAN NON-LINEAR MODELING FOR THE PREDICTION COMPETITION , 1996 .
[79] Alex Fridman,et al. Arguing Machines: Human Supervision of Black Box AI Systems That Make Life-Critical Decisions , 2017, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).