Training Deep Energy-Based Models with f-Divergence Minimization
暂无分享,去创建一个
Lantao Yu | Stefano Ermon | Yang Song | Jiaming Song | S. Ermon | Jiaming Song | Yang Song | Lantao Yu | Stefano Ermon
[1] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[2] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[3] Yee Whye Teh,et al. Bayesian Learning via Stochastic Gradient Langevin Dynamics , 2011, ICML.
[4] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] Prafulla Dhariwal,et al. Glow: Generative Flow with Invertible 1x1 Convolutions , 2018, NeurIPS.
[6] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[7] Jonathon Shlens,et al. Conditional Image Synthesis with Auxiliary Classifier GANs , 2016, ICML.
[8] M. Girolami,et al. Riemann manifold Langevin and Hamiltonian Monte Carlo methods , 2011, Journal of the Royal Statistical Society: Series B (Statistical Methodology).
[9] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[10] Hugo Larochelle,et al. The Neural Autoregressive Distribution Estimator , 2011, AISTATS.
[11] Samy Bengio,et al. Density estimation using Real NVP , 2016, ICLR.
[12] Radford M. Neal. MCMC Using Hamiltonian Dynamics , 2011, 1206.1901.
[13] Thomas P. Minka,et al. Divergence measures and message passing , 2005 .
[14] Hugo Larochelle,et al. MADE: Masked Autoencoder for Distribution Estimation , 2015, ICML.
[15] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[16] Yoshua Bengio,et al. NICE: Non-linear Independent Components Estimation , 2014, ICLR.
[17] Soumith Chintala,et al. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.
[18] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[19] Sebastian Nowozin,et al. Which Training Methods for GANs do actually Converge? , 2018, ICML.
[20] Shakir Mohamed,et al. Learning in Implicit Generative Models , 2016, ArXiv.
[21] Yoshua Bengio,et al. Maximum Entropy Generators for Energy-Based Models , 2019, ArXiv.
[22] Hoon Kim,et al. Monte Carlo Statistical Methods , 2000, Technometrics.
[23] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[24] Martin J. Wainwright,et al. Estimating Divergence Functionals and the Likelihood Ratio by Convex Risk Minimization , 2008, IEEE Transactions on Information Theory.
[25] Aapo Hyvärinen,et al. Noise-Contrastive Estimation of Unnormalized Statistical Models, with Applications to Natural Image Statistics , 2012, J. Mach. Learn. Res..
[26] Sebastian Nowozin,et al. f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.
[27] J. Halton. A Retrospective and Prospective Survey of the Monte Carlo Method , 1970 .
[28] Nicolas Chopin,et al. Noise contrastive estimation: asymptotics, comparison with MC-MLE , 2018, 1801.10381.
[29] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[30] David Barber,et al. Training generative latent models by variational f-divergence minimization , 2018 .
[31] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[32] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[33] Surya Ganguli,et al. Deep Unsupervised Learning using Nonequilibrium Thermodynamics , 2015, ICML.
[34] Pedro M. Domingos,et al. Sum-product networks: A new deep architecture , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).
[35] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[36] Igor Mordatch,et al. Implicit Generation and Modeling with Energy Based Models , 2019, NeurIPS.
[37] Dilin Wang,et al. Learning to Draw Samples: With Application to Amortized MLE for Generative Adversarial Learning , 2016, ArXiv.
[38] Matthias Bethge,et al. A note on the evaluation of generative models , 2015, ICLR.
[39] Erik Nijkamp,et al. Learning Non-Convergent Non-Persistent Short-Run MCMC Toward Energy-Based Model , 2019, NeurIPS.
[40] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[41] Eric Horvitz,et al. Bias Correction of Learned Generative Models using Likelihood-Free Importance Weighting , 2019, DGS@ICLR.
[42] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[43] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[44] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[45] Lantao Yu,et al. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient , 2016, AAAI.
[46] Max Welling,et al. To Relieve Your Headache of Training an MRF, Take AdVIL , 2020, ICLR.
[47] Le Song,et al. Exponential Family Estimation via Adversarial Dynamics Embedding , 2019, NeurIPS.
[48] George Tucker,et al. Energy-Inspired Models: Learning with Sampler-Induced Distributions , 2019, NeurIPS.
[49] J. Zico Kolter,et al. Gradient descent GAN optimization is locally stable , 2017, NIPS.
[50] Alex Graves,et al. Conditional Image Generation with PixelCNN Decoders , 2016, NIPS.
[51] Le Song,et al. Kernel Exponential Family Estimation via Doubly Dual Embedding , 2018, AISTATS.
[52] Tijmen Tieleman,et al. Training restricted Boltzmann machines using approximations to the likelihood gradient , 2008, ICML '08.