暂无分享,去创建一个
Xiang Li | Hui Ye | Shihao Ji | Yang Ye | Xiulong Yang
[1] Richard Zemel,et al. Learning the Stein Discrepancy for Training and Evaluating Energy-Based Models without Sampling , 2020, ICML.
[2] Jun Zhu,et al. Max-Mahalanobis Linear Discriminant Analysis Networks , 2018, ICML.
[3] Erik Nijkamp,et al. On Learning Non-Convergent Short-Run MCMC Toward Energy-Based Model , 2019, ArXiv.
[4] Fu Jie Huang,et al. A Tutorial on Energy-Based Learning , 2006 .
[5] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[6] Xiaojin Zhu,et al. Semi-Supervised Learning , 2010, Encyclopedia of Machine Learning.
[7] Yee Whye Teh,et al. Detecting Out-of-Distribution Inputs to Deep Generative Models Using a Test for Typicality , 2019, ArXiv.
[8] Mohammad Norouzi,et al. Your Classifier is Secretly an Energy Based Model and You Should Treat it Like One , 2019, ICLR.
[9] D. Rubin,et al. Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .
[10] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[11] Yee Whye Teh,et al. Do Deep Generative Models Know What They Don't Know? , 2018, ICLR.
[12] Yang Lu,et al. A Theory of Generative ConvNet , 2016, ICML.
[13] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[14] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[15] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[16] Jiansheng Chen,et al. Rethinking Feature Distribution for Loss Functions in Image Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[17] Andrew M. Dai,et al. Flow Contrastive Estimation of Energy-Based Models , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Kevin Gimpel,et al. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.
[19] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[20] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[21] David Duvenaud,et al. Residual Flows for Invertible Generative Modeling , 2019, NeurIPS.
[22] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[23] Aleksander Madry,et al. Image Synthesis with a Single (Robust) Classifier , 2019, NeurIPS.
[24] H. Robbins. A Stochastic Approximation Method , 1951 .
[25] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[26] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[27] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[29] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[30] Jan Kautz,et al. NVAE: A Deep Hierarchical Variational Autoencoder , 2020, NeurIPS.
[31] David Duvenaud,et al. Invertible Residual Networks , 2018, ICML.
[32] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[33] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[34] Igor Mordatch,et al. Implicit Generation and Generalization with Energy Based Models , 2018 .
[35] Yee Whye Teh,et al. Bayesian Learning via Stochastic Gradient Langevin Dynamics , 2011, ICML.
[36] Jeff Donahue,et al. Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.
[37] Tian Han,et al. On the Anatomy of MCMC-based Maximum Likelihood Learning of Energy-Based Models , 2019, AAAI.
[38] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[39] Ning Chen,et al. Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness , 2019, ICLR.