暂无分享,去创建一个
[1] Yee Whye Teh,et al. Bayesian Learning via Stochastic Gradient Langevin Dynamics , 2011, ICML.
[2] Stephan Günnemann,et al. Posterior Network: Uncertainty Estimation without OOD Samples via Density-Based Pseudo-Counts , 2020, NeurIPS.
[3] Alexander A. Alemi,et al. Density of States Estimation for Out-of-Distribution Detection , 2020, ArXiv.
[4] Kibok Lee,et al. Training Confidence-calibrated Classifiers for Detecting Out-of-Distribution Samples , 2017, ICLR.
[5] Mohammad Norouzi,et al. Your Classifier is Secretly an Energy Based Model and You Should Treat it Like One , 2019, ICLR.
[6] David M. Blei,et al. Prescribed Generative Adversarial Networks , 2019, ArXiv.
[7] S. Mitter,et al. Testing the Manifold Hypothesis , 2013, 1310.0425.
[8] Igor Mordatch,et al. Implicit Generation and Generalization with Energy Based Models , 2018 .
[9] Andrew Gordon Wilson,et al. Why Normalizing Flows Fail to Detect Out-of-Distribution Data , 2020, NeurIPS.
[10] Andrey Malinin,et al. Reverse KL-Divergence Training of Prior Networks: Improved Uncertainty and Adversarial Robustness , 2019, NeurIPS.
[11] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[12] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[13] Hongxia Jin,et al. Generalized ODIN: Detecting Out-of-Distribution Image Without Learning From Out-of-Distribution Data , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Ev Zisselman,et al. Deep Residual Flow for Out of Distribution Detection , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Tonio Ball,et al. Understanding Anomaly Detection with Deep Invertible Networks through Hierarchies of Distributions and Features , 2020, NeurIPS.
[16] Matthias Hein,et al. Why ReLU Networks Yield High-Confidence Predictions Far Away From the Training Data and How to Mitigate the Problem , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Shakir Mohamed,et al. Variational Inference with Normalizing Flows , 2015, ICML.
[18] Eric T. Nalisnick,et al. Detecting Out-of-Distribution Inputs to Deep Generative Models Using Typicality , 2019 .
[19] Yu Cheng,et al. Deep Structured Energy Based Models for Anomaly Detection , 2016, ICML.
[20] Prafulla Dhariwal,et al. Glow: Generative Flow with Invertible 1x1 Convolutions , 2018, NeurIPS.
[21] Kibok Lee,et al. A Simple Unified Framework for Detecting Out-of-Distribution Samples and Adversarial Attacks , 2018, NeurIPS.
[22] Yinda Zhang,et al. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.
[23] Kevin Gimpel,et al. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.
[24] Mohammad Norouzi,et al. No MCMC for me: Amortized sampling for fast and stable training of energy-based models , 2021, ICLR.
[25] Alex Lamb,et al. Deep Learning for Classical Japanese Literature , 2018, ArXiv.
[26] Jordi Luque,et al. Input complexity and out-of-distribution detection with likelihood-based generative models , 2020, ICLR.
[27] Michael U. Gutmann,et al. Conditional Noise-Contrastive Estimation of Unnormalised Models , 2018, ICML.
[28] Erik Nijkamp,et al. Learning Non-Convergent Non-Persistent Short-Run MCMC Toward Energy-Based Model , 2019, NeurIPS.
[29] Francisco J. R. Ruiz,et al. Unbiased Implicit Variational Inference , 2018, AISTATS.
[30] Simon Haykin,et al. GradientBased Learning Applied to Document Recognition , 2001 .
[31] Andrew M. Dai,et al. Flow Contrastive Estimation of Energy-Based Models , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[33] Aapo Hyvärinen,et al. Estimation of Non-Normalized Statistical Models by Score Matching , 2005, J. Mach. Learn. Res..
[34] Yang Song,et al. Sliced Score Matching: A Scalable Approach to Density and Score Estimation , 2019, UAI.
[35] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[36] Kui Yuan,et al. A Compact Convolutional Neural Network for Surface Defect Inspection , 2020, Sensors.
[37] Yee Whye Teh,et al. Do Deep Generative Models Know What They Don't Know? , 2018, ICLR.
[38] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[39] Weitang Liu,et al. Energy-based Out-of-distribution Detection , 2020, NeurIPS.
[40] Jasper Snoek,et al. Likelihood Ratios for Out-of-Distribution Detection , 2019, NeurIPS.
[41] Kush R. Varshney,et al. Engineering safety in machine learning , 2016, 2016 Information Theory and Applications Workshop (ITA).
[42] Kumar Sricharan,et al. Building robust classifiers through generation of confident out of distribution examples , 2018, ArXiv.
[43] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[44] Tijmen Tieleman,et al. Training restricted Boltzmann machines using approximations to the likelihood gradient , 2008, ICML '08.
[45] Charles Blundell,et al. Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles , 2016, NIPS.
[46] R. Srikant,et al. Enhancing The Reliability of Out-of-distribution Image Detection in Neural Networks , 2017, ICLR.
[47] Thomas G. Dietterich,et al. Deep Anomaly Detection with Outlier Exposure , 2018, ICLR.
[48] Aapo Hyvärinen,et al. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models , 2010, AISTATS.
[49] Alexander A. Alemi,et al. WAIC, but Why? Generative Ensembles for Robust Anomaly Detection , 2018 .
[50] Mark J. F. Gales,et al. Predictive Uncertainty Estimation via Prior Networks , 2018, NeurIPS.