暂无分享,去创建一个
[1] Pieter Abbeel,et al. PixelSNAIL: An Improved Autoregressive Generative Model , 2017, ICML.
[2] Yang Lu,et al. Learning Generative ConvNets via Multi-grid Modeling and Sampling , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[3] Yang Song,et al. Generative Modeling by Estimating Gradients of the Data Distribution , 2019, NeurIPS.
[4] Surya Ganguli,et al. Deep Unsupervised Learning using Nonequilibrium Thermodynamics , 2015, ICML.
[5] Emiel Hoogeboom,et al. Predictive Sampling with Forecasting Autoregressive Models , 2020, ICML.
[6] Alex Nichol. VQ-DRAW: A Sequential Discrete VAE , 2020, ArXiv.
[7] Song-Chun Zhu,et al. Synthesizing Dynamic Patterns by Spatial-Temporal Generative ConvNet , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Rémi Munos,et al. Autoregressive Quantile Networks for Generative Modeling , 2018, ICML.
[9] Igor Mordatch,et al. Implicit Generation and Modeling with Energy Based Models , 2019, NeurIPS.
[10] Song-Chun Zhu,et al. Learning Descriptor Networks for 3D Shape Synthesis and Analysis , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[11] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[12] Ilya Sutskever,et al. Generating Long Sequences with Sparse Transformers , 2019, ArXiv.
[13] Hao Wu,et al. Stochastic Normalizing Flows , 2020, NeurIPS.
[14] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[15] Nal Kalchbrenner,et al. Generating High Fidelity Images with Subscale Pixel Networks and Multidimensional Upscaling , 2018, ICLR.
[16] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[17] Pieter Abbeel,et al. Flow++: Improving Flow-Based Generative Models with Variational Dequantization and Architecture Design , 2019, ICML.
[18] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[19] Jaakko Lehtinen,et al. Training Generative Adversarial Networks with Limited Data , 2020, NeurIPS.
[20] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[21] Sina Honari,et al. Learning to Generate Samples from Noise through Infusion Training , 2017, ICLR.
[22] Andrew M. Dai,et al. Flow Contrastive Estimation of Energy-Based Models , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Abhinav Gupta,et al. Non-local Neural Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[24] Robert Peharz,et al. Minimal Random Code Learning: Getting Bits Back from Compressed Model Parameters , 2018, ICLR.
[25] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[26] Max Welling,et al. Markov Chain Monte Carlo and Variational Inference: Bridging the Gap , 2014, ICML.
[27] Jascha Sohl-Dickstein,et al. Generalizing Hamiltonian Monte Carlo with Neural Networks , 2017, ICLR.
[28] Ryan Prenger,et al. Waveglow: A Flow-based Generative Network for Speech Synthesis , 2018, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[29] Ali Razavi,et al. Generating Diverse High-Fidelity Images with VQ-VAE-2 , 2019, NeurIPS.
[30] Yang Lu,et al. A Theory of Generative ConvNet , 2016, ICML.
[31] Jaakko Lehtinen,et al. Progressive Growing of GANs for Improved Quality, Stability, and Variation , 2017, ICLR.
[32] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[33] David Duvenaud,et al. Neural Ordinary Differential Equations , 2018, NeurIPS.
[34] Max Welling,et al. Improved Variational Inference with Inverse Autoregressive Flow , 2016, NIPS 2016.
[35] Jeff Donahue,et al. Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.
[36] Tian Han,et al. On the Anatomy of MCMC-based Maximum Likelihood Learning of Energy-Based Models , 2019, AAAI.
[37] Song-Chun Zhu,et al. Learning Energy-Based Spatial-Temporal Generative ConvNets for Dynamic Patterns , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[38] Myle Ott,et al. Residual Energy-Based Models for Text Generation , 2020, ICLR.
[39] Shakir Mohamed,et al. Variational Inference with Normalizing Flows , 2015, ICML.
[40] Ole Winther,et al. BIVA: A Very Deep Hierarchy of Latent Variables for Generative Modeling , 2019, NeurIPS.
[41] Andrew Owens,et al. CNN-Generated Images Are Surprisingly Easy to Spot… for Now , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[42] Xi Chen,et al. PixelCNN++: Improving the PixelCNN with Discretized Logistic Mixture Likelihood and Other Modifications , 2017, ICLR.
[43] George Tucker,et al. Energy-Inspired Models: Learning with Sampler-Induced Distributions , 2019, NeurIPS.
[44] Kaiming He,et al. Group Normalization , 2018, ECCV.
[45] Yoshua Bengio,et al. NICE: Non-linear Independent Components Estimation , 2014, ICLR.
[46] Alex Graves,et al. Conditional Image Generation with PixelCNN Decoders , 2016, NIPS.
[47] Alireza Makhzani,et al. Evaluating Lossy Compression Rates of Deep Generative Models , 2020, ICML.
[48] Timo Aila,et al. A Style-Based Generator Architecture for Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[49] Surya Ganguli,et al. Variational Walkback: Learning a Transition Operator as a Stochastic Recurrent Net , 2017, NIPS.
[50] Pascal Vincent,et al. A Connection Between Score Matching and Denoising Autoencoders , 2011, Neural Computation.
[51] Yinda Zhang,et al. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.
[52] Prafulla Dhariwal,et al. Glow: Generative Flow with Invertible 1x1 Convolutions , 2018, NeurIPS.
[53] Samy Bengio,et al. Density estimation using Real NVP , 2016, ICLR.
[54] Erich Elsen,et al. Efficient Neural Audio Synthesis , 2018, ICML.
[55] Alex Graves,et al. Video Pixel Networks , 2016, ICML.
[56] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[57] Jaikumar Radhakrishnan,et al. The Communication Complexity of Correlation , 2007, IEEE Transactions on Information Theory.
[58] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[59] Jaakko Lehtinen,et al. Analyzing and Improving the Image Quality of StyleGAN , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[60] Heiga Zen,et al. WaveNet: A Generative Model for Raw Audio , 2016, SSW.
[61] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[62] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[63] Tim Salimans,et al. Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks , 2016, NIPS.
[64] Tong Che,et al. Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling , 2020, NeurIPS.
[65] Mohammad Norouzi,et al. Your Classifier is Secretly an Energy Based Model and You Should Treat it Like One , 2019, ICLR.
[66] Thomas Brox,et al. U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.
[67] Daan Wierstra,et al. Towards Conceptual Compression , 2016, NIPS.
[68] Stefano Ermon,et al. A-NICE-MC: Adversarial Training for MCMC , 2017, NIPS.
[69] Stefano Ermon,et al. Improved Techniques for Training Score-Based Generative Models , 2020, NeurIPS.
[70] David Duvenaud,et al. FFJORD: Free-form Continuous Dynamics for Scalable Reversible Generative Models , 2018, ICLR.
[71] Erik Nijkamp,et al. Learning Non-Convergent Non-Persistent Short-Run MCMC Toward Energy-Based Model , 2019, NeurIPS.