Alleviation of Gradient Exploding in GANs: Fake Can Be Real
暂无分享,去创建一个
[1] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[2] Yann LeCun,et al. Energy-based Generative Adversarial Network , 2016, ICLR.
[3] Yingyu Liang,et al. Generalization and Equilibrium in Generative Adversarial Nets (GANs) , 2017, ICML.
[4] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[5] Trung Le,et al. MGAN: Training Generative Adversarial Nets with Multiple Generators , 2018, ICLR.
[6] Han Zhang,et al. Self-Attention Generative Adversarial Networks , 2018, ICML.
[7] Stefan Winkler,et al. The Unusual Effectiveness of Averaging in GAN Training , 2018, ICLR.
[8] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[9] Ting Chen,et al. On Self Modulation for Generative Adversarial Networks , 2018, ICLR.
[10] Philip H. S. Torr,et al. Multi-agent Diverse Generative Adversarial Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[11] Jaakko Lehtinen,et al. Progressive Growing of GANs for Improved Quality, Stability, and Variation , 2017, ICLR.
[12] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[13] Jeff Donahue,et al. Large Scale Adversarial Representation Learning , 2019, NeurIPS.
[14] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[15] Denis Lukovnikov,et al. On the regularization of Wasserstein GANs , 2017, ICLR.
[16] Ashish Khetan,et al. PacGAN: The Power of Two Samples in Generative Adversarial Networks , 2017, IEEE Journal on Selected Areas in Information Theory.
[17] David M. Blei,et al. Prescribed Generative Adversarial Networks , 2019, ArXiv.
[18] Guo-Jun Qi,et al. Loss-Sensitive Generative Adversarial Networks on Lipschitz Densities , 2017, International Journal of Computer Vision.
[19] Soumith Chintala,et al. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.
[20] Jun Zhou,et al. Generalization in Generative Adversarial Networks: A Novel Perspective from Privacy Protection , 2019, NeurIPS.
[21] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[22] Sebastian Nowozin,et al. Which Training Methods for GANs do actually Converge? , 2018, ICML.
[23] J. Zico Kolter,et al. Gradient descent GAN optimization is locally stable , 2017, NIPS.
[24] Timo Aila,et al. A Style-Based Generator Architecture for Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Antonio Torralba,et al. Ieee Transactions on Pattern Analysis and Machine Intelligence 1 80 Million Tiny Images: a Large Dataset for Non-parametric Object and Scene Recognition , 2022 .
[26] Yi Zhang,et al. Do GANs learn the distribution? Some Theory and Empirics , 2018, ICLR.
[27] Truyen Tran,et al. Improving Generalization and Stability of Generative Adversarial Networks , 2019, ICLR.
[28] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[29] Jeff Donahue,et al. Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.
[30] Yann Ollivier,et al. Mixed batches and symmetric discriminators for GAN training , 2018, ICML.
[31] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[32] Sebastian Nowozin,et al. f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.
[33] Sebastian Nowozin,et al. Stabilizing Training of Generative Adversarial Networks through Regularization , 2017, NIPS.
[34] Guang-He Lee,et al. Bayesian Modelling and Monte Carlo Inference for GAN , 2018 .
[35] Raymond Y. K. Lau,et al. Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).