暂无分享,去创建一个
[1] Léon Bottou,et al. Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.
[2] Ole Winther,et al. Autoencoding beyond pixels using a learned similarity metric , 2015, ICML.
[3] Aaron C. Courville,et al. Adversarially Learned Inference , 2016, ICLR.
[4] Sebastian Nowozin,et al. Stabilizing Training of Generative Adversarial Networks through Regularization , 2017, NIPS.
[5] Ngai-Man Cheung,et al. Efficient and Deep Person Re-identification Using Multi-level Similarity , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[6] Navdeep Jaitly,et al. Adversarial Autoencoders , 2015, ArXiv.
[7] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[8] J. Zico Kolter,et al. Gradient descent GAN optimization is locally stable , 2017, NIPS.
[9] Stefan Winkler,et al. The Unusual Effectiveness of Averaging in GAN Training , 2018, ICLR.
[10] Yuval Elovici,et al. DOPING: Generative Data Augmentation for Unsupervised Anomaly Detection with GAN , 2018, 2018 IEEE International Conference on Data Mining (ICDM).
[11] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[12] Denis Lukovnikov,et al. On the regularization of Wasserstein GANs , 2017, ICLR.
[13] Trevor Darrell,et al. Adversarial Feature Learning , 2016, ICLR.
[14] Sebastian Nowozin,et al. The Numerics of GANs , 2017, NIPS.
[15] Sebastian Nowozin,et al. f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.
[16] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[17] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[18] Truyen Tran,et al. On catastrophic forgetting and mode collapse in Generative Adversarial Networks , 2018, ArXiv.
[19] Marc G. Bellemare,et al. The Cramer Distance as a Solution to Biased Wasserstein Gradients , 2017, ArXiv.
[20] Yann LeCun,et al. Energy-based Generative Adversarial Network , 2016, ICLR.
[21] Kanglin Liu. Varying k-Lipschitz Constraint for Generative Adversarial Networks , 2018, ArXiv.
[22] David Berthelot,et al. BEGAN: Boundary Equilibrium Generative Adversarial Networks , 2017, ArXiv.
[23] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[24] Yoshua Bengio,et al. Mode Regularized Generative Adversarial Networks , 2016, ICLR.
[25] Ian J. Goodfellow,et al. NIPS 2016 Tutorial: Generative Adversarial Networks , 2016, ArXiv.
[26] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[27] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[28] Tanmoy Chakraborty,et al. MMGAN: Manifold Matching Generative Adversarial Network for Generating Images , 2017, ArXiv.
[29] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[30] David Pfau,et al. Unrolled Generative Adversarial Networks , 2016, ICLR.
[31] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[32] Ngai-Man Cheung,et al. Dist-GAN: An Improved GAN Using Distance Constraints , 2018, ECCV.
[33] Jacob Abernethy,et al. On Convergence and Stability of GANs , 2018 .
[34] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[35] Sebastian Nowozin,et al. Which Training Methods for GANs do actually Converge? , 2018, ICML.