HOW WELL GENERATIVE ADVERSARIAL NETWORKS LEARN DISTRIBUTIONS1
暂无分享,去创建一个
[1] C. J. Stone,et al. Optimal Global Rates of Convergence for Nonparametric Regression , 1982 .
[2] L. Hansen. Large Sample Properties of Generalized Method of Moments Estimators , 1982 .
[3] D. Pollard,et al. Simulation and the Asymptotics of Optimization Estimators , 1989 .
[4] D. McFadden. A Method of Simulated Moments for Estimation of Discrete Response Models Without Numerical Integration , 1989 .
[5] D. Pollard. Empirical Processes: Theory and Applications , 1990 .
[6] L. Caffarelli. Some regularity properties of solutions of Monge Ampère equation , 1991 .
[7] L. Caffarelli. The regularity of mappings with a convex potential , 1992 .
[8] K. Back,et al. Implied Probabilities in GMM Estimators , 1993 .
[9] G. Imbens,et al. Information Theoretic Approaches to Inference in Moment Condition Models , 1995 .
[10] Bernard A. Mair,et al. Statistical Inverse Estimation in Hilbert Scales , 1996, SIAM J. Appl. Math..
[11] Peter L. Bartlett,et al. Neural Network Learning - Theoretical Foundations , 1999 .
[12] R. Nickl,et al. Bracketing Metric Entropy Rates and Empirical Central Limit Theorems for Function Classes of Besov- and Sobolev-Type , 2007 .
[13] A. Caponnetto,et al. Optimal Rates for the Regularized Least-Squares Algorithm , 2007, Found. Comput. Math..
[14] Alexandre B. Tsybakov,et al. Introduction to Nonparametric Estimation , 2008, Springer series in statistics.
[15] Lorenzo Rosasco,et al. Learning Probability Measures with respect to Optimal Transport Metrics , 2012, NIPS.
[16] L. Ambrosio,et al. A User’s Guide to Optimal Transport , 2013 .
[17] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[18] Zoubin Ghahramani,et al. Training generative neural networks via Maximum Mean Discrepancy optimization , 2015, UAI.
[19] Tengyuan Liang,et al. Law of log determinant of sample covariance matrix and optimal estimation of differential entropy for high-dimensional Gaussian distributions , 2013, Journal of Multivariate Analysis.
[20] Richard S. Zemel,et al. Generative Moment Matching Networks , 2015, ICML.
[21] Tengyuan Liang,et al. How Well Can Generative Adversarial Networks (GAN) Learn Densities: A Nonparametric View , 2017, ArXiv.
[22] Lucas Theis,et al. Amortised MAP Inference for Image Super-resolution , 2016, ICLR.
[23] Sebastian Nowozin,et al. The Numerics of GANs , 2017, NIPS.
[24] Dmitry Yarotsky,et al. Error bounds for approximations with deep ReLU networks , 2016, Neural Networks.
[25] Yingyu Liang,et al. Generalization and Equilibrium in Generative Adversarial Nets (GANs) , 2017, ICML.
[26] Abbas Mehrabian,et al. Nearly-tight VC-dimension bounds for piecewise linear neural networks , 2017, COLT.
[27] Léon Bottou,et al. Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.
[28] Kamalika Chaudhuri,et al. Approximation and Convergence Properties of Generative Adversarial Learning , 2017, NIPS.
[29] Yi Zhang,et al. Do GANs actually learn the distribution? An empirical study , 2017, ArXiv.
[30] Arthur Gretton,et al. On gradient regularizers for MMD GANs , 2018, NeurIPS.
[31] Kamalika Chaudhuri,et al. The Inductive Bias of Restricted f-GANs , 2018, ArXiv.
[32] Constantinos Daskalakis,et al. Training GANs with Optimism , 2017, ICLR.
[33] Sebastian Nowozin,et al. Which Training Methods for GANs do actually Converge? , 2018, ICML.
[34] Mario Lucic,et al. Are GANs Created Equal? A Large-Scale Study , 2017, NeurIPS.
[35] Guido Imbens,et al. Using Wasserstein Generative Adversarial Networks for the Design of Monte Carlo Simulations , 2019, Journal of Econometrics.
[36] Tengyuan Liang,et al. Interaction Matters: A Note on Non-asymptotic Local Convergence of Generative Adversarial Networks , 2018, AISTATS.
[37] Tengyuan Liang. Estimating Certain Integral Probability Metrics (IPMs) Is as Hard as Estimating under the IPMs , 2019, SSRN Electronic Journal.