HOW WELL GENERATIVE ADVERSARIAL NETWORKS LEARN DISTRIBUTIONS1

This paper studies the rates of convergence for learning distributions implicitly with the adversarial framework and Generative Adversarial Networks (GAN), which subsume Wasserstein, Sobolev, MMD GAN, and Generalized/Simulated Method of Moments (GMM/SMM) as special cases. We study a wide range of parametric and nonparametric target distributions, under a host of objective evaluation metrics. We investigate how to obtain a good statistical guarantee for GANs through the lens of regularization. On the nonparametric end, we derive the optimal minimax rates for distribution estimation under the adversarial framework. On the parametric end, we establish a theory for general neural network classes (including deep leaky ReLU networks), that characterizes the interplay on the choice of generator and discriminator pair. We discover and isolate a new notion of regularization, called the generator-discriminator-pair regularization, that sheds light on the advantage of GANs compared to classical parametric and nonparametric approaches for explicit distribution estimation. We develop novel oracle inequalities as the main technical tools for analyzing GANs, which is of independent interest.

[1]  C. J. Stone,et al.  Optimal Global Rates of Convergence for Nonparametric Regression , 1982 .

[2]  L. Hansen Large Sample Properties of Generalized Method of Moments Estimators , 1982 .

[3]  D. Pollard,et al.  Simulation and the Asymptotics of Optimization Estimators , 1989 .

[4]  D. McFadden A Method of Simulated Moments for Estimation of Discrete Response Models Without Numerical Integration , 1989 .

[5]  D. Pollard Empirical Processes: Theory and Applications , 1990 .

[6]  L. Caffarelli Some regularity properties of solutions of Monge Ampère equation , 1991 .

[7]  L. Caffarelli The regularity of mappings with a convex potential , 1992 .

[8]  K. Back,et al.  Implied Probabilities in GMM Estimators , 1993 .

[9]  G. Imbens,et al.  Information Theoretic Approaches to Inference in Moment Condition Models , 1995 .

[10]  Bernard A. Mair,et al.  Statistical Inverse Estimation in Hilbert Scales , 1996, SIAM J. Appl. Math..

[11]  Peter L. Bartlett,et al.  Neural Network Learning - Theoretical Foundations , 1999 .

[12]  R. Nickl,et al.  Bracketing Metric Entropy Rates and Empirical Central Limit Theorems for Function Classes of Besov- and Sobolev-Type , 2007 .

[13]  A. Caponnetto,et al.  Optimal Rates for the Regularized Least-Squares Algorithm , 2007, Found. Comput. Math..

[14]  Alexandre B. Tsybakov,et al.  Introduction to Nonparametric Estimation , 2008, Springer series in statistics.

[15]  Lorenzo Rosasco,et al.  Learning Probability Measures with respect to Optimal Transport Metrics , 2012, NIPS.

[16]  L. Ambrosio,et al.  A User’s Guide to Optimal Transport , 2013 .

[17]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[18]  Zoubin Ghahramani,et al.  Training generative neural networks via Maximum Mean Discrepancy optimization , 2015, UAI.

[19]  Tengyuan Liang,et al.  Law of log determinant of sample covariance matrix and optimal estimation of differential entropy for high-dimensional Gaussian distributions , 2013, Journal of Multivariate Analysis.

[20]  Richard S. Zemel,et al.  Generative Moment Matching Networks , 2015, ICML.

[21]  Tengyuan Liang,et al.  How Well Can Generative Adversarial Networks (GAN) Learn Densities: A Nonparametric View , 2017, ArXiv.

[22]  Lucas Theis,et al.  Amortised MAP Inference for Image Super-resolution , 2016, ICLR.

[23]  Sebastian Nowozin,et al.  The Numerics of GANs , 2017, NIPS.

[24]  Dmitry Yarotsky,et al.  Error bounds for approximations with deep ReLU networks , 2016, Neural Networks.

[25]  Yingyu Liang,et al.  Generalization and Equilibrium in Generative Adversarial Nets (GANs) , 2017, ICML.

[26]  Abbas Mehrabian,et al.  Nearly-tight VC-dimension bounds for piecewise linear neural networks , 2017, COLT.

[27]  Léon Bottou,et al.  Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.

[28]  Kamalika Chaudhuri,et al.  Approximation and Convergence Properties of Generative Adversarial Learning , 2017, NIPS.

[29]  Yi Zhang,et al.  Do GANs actually learn the distribution? An empirical study , 2017, ArXiv.

[30]  Arthur Gretton,et al.  On gradient regularizers for MMD GANs , 2018, NeurIPS.

[31]  Kamalika Chaudhuri,et al.  The Inductive Bias of Restricted f-GANs , 2018, ArXiv.

[32]  Constantinos Daskalakis,et al.  Training GANs with Optimism , 2017, ICLR.

[33]  Sebastian Nowozin,et al.  Which Training Methods for GANs do actually Converge? , 2018, ICML.

[34]  Mario Lucic,et al.  Are GANs Created Equal? A Large-Scale Study , 2017, NeurIPS.

[35]  Guido Imbens,et al.  Using Wasserstein Generative Adversarial Networks for the Design of Monte Carlo Simulations , 2019, Journal of Econometrics.

[36]  Tengyuan Liang,et al.  Interaction Matters: A Note on Non-asymptotic Local Convergence of Generative Adversarial Networks , 2018, AISTATS.

[37]  Tengyuan Liang Estimating Certain Integral Probability Metrics (IPMs) Is as Hard as Estimating under the IPMs , 2019, SSRN Electronic Journal.