暂无分享,去创建一个
[1] Yarin Gal,et al. Uncertainty in Deep Learning , 2016 .
[2] O. Macchi. The coincidence approach to stochastic point processes , 1975, Advances in Applied Probability.
[3] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[4] Shin-ichi Maeda,et al. A Bayesian encourages dropout , 2014, ArXiv.
[5] Klaus C. J. Dietmayer,et al. Towards Safe Autonomous Driving: Capture Uncertainty in the Deep Neural Network For Lidar 3D Vehicle Detection , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).
[6] Kevin Gimpel,et al. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.
[7] Linton G. Freeman,et al. Elementary Applied Statistics for students in Behavioral Science , 1965 .
[8] Xia Zhu,et al. Out-of-Distribution Detection Using an Ensemble of Self Supervised Leave-out Classifiers , 2018, ECCV.
[9] S. Goreinov,et al. How to find a good submatrix , 2010 .
[10] Jonas Mueller,et al. Maximizing Overall Diversity for Improved Uncertainty Estimates in Deep Ensembles , 2019, AAAI.
[11] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[12] Jorge J. Moré,et al. Digital Object Identifier (DOI) 10.1007/s101070100263 , 2001 .
[13] Malik Magdon-Ismail,et al. On selecting a maximum volume sub-matrix of a matrix and related problems , 2009, Theor. Comput. Sci..
[14] Andrew Gordon Wilson,et al. A Simple Baseline for Bayesian Uncertainty in Deep Learning , 2019, NeurIPS.
[15] Andrew Gordon Wilson,et al. Subspace Inference for Bayesian Deep Learning , 2019, UAI.
[16] Matthias Hein,et al. Why ReLU Networks Yield High-Confidence Predictions Far Away From the Training Data and How to Mitigate the Problem , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Ryan P. Adams,et al. Probabilistic Backpropagation for Scalable Learning of Bayesian Neural Networks , 2015, ICML.
[18] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[19] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[20] Evgenii Tsymbalov,et al. Dropout-based Active Learning for Regression , 2018, AIST.
[21] Sven Wåhlin,et al. [Who can you trust?]. , 2012, Lakartidningen.
[22] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[23] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[24] Andrew Gordon Wilson,et al. Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs , 2018, NeurIPS.
[25] Ben Taskar,et al. Determinantal Point Processes for Machine Learning , 2012, Found. Trends Mach. Learn..
[26] Michal Valko,et al. DPPy: DPP Sampling with Python , 2019, J. Mach. Learn. Res..
[27] Andreas Nürnberger,et al. The Power of Ensembles for Active Learning in Image Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[28] Adrian E. Roitberg,et al. Less is more: sampling chemical space with active learning , 2018, The Journal of chemical physics.
[29] Zoubin Ghahramani,et al. Bayesian Active Learning for Classification and Preference Learning , 2011, ArXiv.
[30] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[31] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[32] Zoubin Ghahramani,et al. Deep Bayesian Active Learning with Image Data , 2017, ICML.
[33] Sang Joon Kim,et al. A Mathematical Theory of Communication , 2006 .
[34] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[35] Tanmoy Bhattacharya,et al. The need for uncertainty quantification in machine-assisted medical decision making , 2019, Nat. Mach. Intell..
[36] Padhraic Smyth,et al. Dropout as a Structured Shrinkage Prior , 2018, ICML.
[37] David Cohn,et al. Active Learning , 2010, Encyclopedia of Machine Learning.