Implicit competitive regularization in GANs
暂无分享,去创建一个
Anima Anandkumar | Florian Schäfer | Hongkai Zheng | Anima Anandkumar | Florian Schäfer | Hongkai Zheng
[1] Thore Graepel,et al. Differentiable Game Mechanics , 2019, J. Mach. Learn. Res..
[2] Masayoshi Kubo,et al. Implicit Regularization in Over-parameterized Neural Networks , 2019, ArXiv.
[3] Sebastian Nowozin,et al. The Numerics of GANs , 2017, NIPS.
[4] Shimon Whiteson,et al. Learning with Opponent-Learning Awareness , 2017, AAMAS.
[5] Yuxin Chen,et al. Implicit Regularization in Nonconvex Statistical Estimation: Gradient Descent Converges Linearly for Phase Retrieval, Matrix Completion, and Blind Deconvolution , 2017, Found. Comput. Math..
[6] Gauthier Gidel,et al. Parametric Adversarial Divergences are Good Task Losses for Generative Modeling , 2017, ICLR.
[7] Nathan Srebro,et al. Implicit Regularization in Matrix Factorization , 2017, 2018 Information Theory and Applications Workshop (ITA).
[8] Michael I. Jordan,et al. What is Local Optimality in Nonconvex-Nonconcave Minimax Optimization? , 2019, ICML.
[9] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..
[10] Kenji Fukumizu,et al. On integral probability metrics, φ-divergences and binary classification , 2009, 0901.2698.
[11] Constantinos Daskalakis,et al. Training GANs with Optimism , 2017, ICLR.
[12] Constantinos Daskalakis,et al. The Limit Points of (Optimistic) Gradient Descent in Min-Max Optimization , 2018, NeurIPS.
[13] Babak Hassibi,et al. Stochastic Mirror Descent on Overparameterized Nonlinear Models , 2019, IEEE Transactions on Neural Networks and Learning Systems.
[14] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[15] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[16] S. Shankar Sastry,et al. On Gradient-Based Learning in Continuous Games , 2018, SIAM J. Math. Data Sci..
[17] Alain Trouvé,et al. Metamorphoses Through Lie Group Action , 2005, Found. Comput. Math..
[18] Jonas Adler,et al. Banach Wasserstein GAN , 2018, NeurIPS.
[19] Arthur Gretton,et al. Demystifying MMD GANs , 2018, ICLR.
[20] Michael I. Jordan,et al. Minmax Optimization: Stable Limit Points of Gradient Descent Ascent are Locally Optimal , 2019, ArXiv.
[21] Raymond Y. K. Lau,et al. Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[22] Hans Burkhardt,et al. Invariant kernel functions for pattern analysis and machine learning , 2007, Machine Learning.
[23] Satrajit Chatterjee,et al. Coherent Gradients: An Approach to Understanding Generalization in Gradient Descent-based Optimization , 2020, ICLR.
[24] Pascal Vincent,et al. A Closer Look at the Optimization Landscapes of Generative Adversarial Networks , 2019, ICLR.
[25] Lillian J. Ratliff,et al. Convergence of Learning Dynamics in Stackelberg Games , 2019, ArXiv.
[26] Richard S. Zemel,et al. Generative Moment Matching Networks , 2015, ICML.
[27] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[28] Florian Schäfer,et al. Image Extrapolation for the Time Discrete Metamorphosis Model: Existence and Applications , 2017, SIAM J. Imaging Sci..
[29] Sanjeev Arora,et al. Implicit Regularization in Deep Matrix Factorization , 2019, NeurIPS.
[30] G. M. Korpelevich. The extragradient method for finding saddle points and other problems , 1976 .
[31] Sridhar Mahadevan,et al. Global Convergence to the Equilibrium of GANs using Variational Inequalities , 2018, ArXiv.
[32] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[33] Mario Lucic,et al. Are GANs Created Equal? A Large-Scale Study , 2017, NeurIPS.
[34] Bill Triggs,et al. Histograms of oriented gradients for human detection , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).
[35] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[36] Yiming Yang,et al. MMD GAN: Towards Deeper Understanding of Moment Matching Network , 2017, NIPS.
[37] Yann LeCun,et al. Transformation Invariance in Pattern Recognition-Tangent Distance and Tangent Propagation , 1996, Neural Networks: Tricks of the Trade.
[38] Lirong Dai,et al. Local Coding Based Matching Kernel Method for Image Classification , 2014, PloS one.
[39] Benjamin Berkels,et al. Time Discrete Geodesic Paths in the Space of Images , 2015, SIAM J. Imaging Sci..
[40] Behnam Neyshabur,et al. Implicit Regularization in Deep Learning , 2017, ArXiv.
[41] Thore Graepel,et al. The Mechanics of n-Player Differentiable Games , 2018, ICML.
[42] V. Tikhomirov. On the Representation of Continuous Functions of Several Variables as Superpositions of Continuous Functions of a Smaller Number of Variables , 1991 .
[43] S. Shankar Sastry,et al. On Finding Local Nash Equilibria (and Only Local Nash Equilibria) in Zero-Sum Games , 2019, 1901.00838.
[44] Sebastian Nowozin,et al. f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.
[45] Léon Bottou,et al. Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.
[46] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[47] Daan Wierstra,et al. Stochastic Backpropagation and Approximate Inference in Deep Generative Models , 2014, ICML.
[48] Jacob Abernethy,et al. On Convergence and Stability of GANs , 2018 .
[49] Yu Cheng,et al. Sobolev GAN , 2017, ICLR.
[50] Yingyu Liang,et al. Generalization and Equilibrium in Generative Adversarial Nets (GANs) , 2017, ICML.
[51] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[52] Sebastian Nowozin,et al. Stabilizing Training of Generative Adversarial Networks through Regularization , 2017, NIPS.
[53] Jian Peng,et al. A Gradual, Semi-Discrete Approach to Generative Network Training via Explicit Wasserstein Minimization , 2019, ICML.
[54] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[55] S. Eisenstat. Efficient Implementation of a Class of Preconditioned Conjugate Gradient Methods , 1981 .
[56] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[57] Florian Schäfer,et al. Competitive Gradient Descent , 2019, NeurIPS.