A Deep Generative Approach to Conditional Sampling

We propose a deep generative approach to sampling from a conditional distribution based on a unified formulation of conditional distribution and generalized nonparametric regression function using the noise-outsourcing lemma. The proposed approach aims at learning a conditional generator so that a random sample from the target conditional distribution can be obtained by the action of the conditional generator on a sample drawn from a reference distribution. The conditional generator is estimated nonparametrically with neural networks by matching appropriate joint distributions using Kullback-Liebler divergence. An appealing aspect of our method is that it allows either of or both the predictor and the response to be high-dimensional and can handle both continuous and discrete type predictors and responses. We show that the proposed method is consistent in the sense that the conditional generator converges in distribution to the underlying conditional distribution under mild conditions. Our numerical experiments with simulated and benchmark image data validate the proposed method and demonstrate that it outperforms several existing conditional density estimation methods.

[1]  S. M. Ali,et al.  A General Class of Coefficients of Divergence of One Distribution from Another , 1966 .

[2]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[3]  D. W. Scott,et al.  Multivariate Density Estimation, Theory, Practice and Visualization , 1992 .

[4]  Peter L. Bartlett,et al.  Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..

[5]  O. Kallenberg Foundations of Modern Probability , 2021, Probability Theory and Stochastic Modelling.

[6]  Alexandre B. Tsybakov,et al.  Introduction to Nonparametric Estimation , 2008, Springer series in statistics.

[7]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[8]  Qiwei Yao,et al.  Approximating conditional distribution functions using dimension reduction , 2005 .

[9]  Jeffrey S. Racine,et al.  Cross-Validation and the Estimation of Conditional Probability Densities , 2004 .

[10]  Sebastian Nowozin,et al.  f-GAN: Training Generative Neural Samplers using Variational Divergence Minimization , 2016, NIPS.

[11]  Zuowei Shen,et al.  Deep Network Approximation Characterized by Number of Neurons , 2019, Communications in Computational Physics.

[12]  Rob J Hyndman,et al.  Estimating and Visualizing Conditional Densities , 1996 .

[13]  R. Cook Regression Graphics , 1994 .

[14]  M. Kohler,et al.  On deep learning as a remedy for the curse of dimensionality in nonparametric regression , 2019, The Annals of Statistics.

[15]  丸山 徹 Convex Analysisの二,三の進展について , 1977 .

[16]  David W. Scott,et al.  Multivariate Density Estimation: Theory, Practice, and Visualization , 1992, Wiley Series in Probability and Statistics.

[17]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[18]  Tim Austin,et al.  Exchangeable random measures , 2013, 1302.2116.

[19]  Michael Kohler,et al.  Nonparametric estimation of a conditional density , 2017 .

[20]  R. H. Moore,et al.  Regression Graphics: Ideas for Studying Regressions Through Graphics , 1998, Technometrics.

[21]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[22]  Marie Frei,et al.  Decoupling From Dependence To Independence , 2016 .

[23]  Michael Kohler,et al.  On the rate of convergence of fully connected very deep neural network regression estimates , 2019, The Annals of Statistics.

[24]  Yuling Jiao,et al.  Robust Nonparametric Regression with Deep Neural Networks , 2021, 2107.10343.

[25]  Rafael Izbicki,et al.  Converting High-Dimensional Regression to High-Dimensional Conditional Density Estimation , 2017, 1704.08095.

[26]  Jon A. Wellner,et al.  Weak Convergence and Empirical Processes: With Applications to Statistics , 1996 .

[27]  Rafael Izbicki,et al.  Conditional Density Estimation Tools in Python and R with Applications to Photometric Redshifts and Likelihood-Free Cosmological Inference , 2019 .

[28]  Martin J. Wainwright,et al.  Estimating Divergence Functionals and the Likelihood Ratio by Convex Risk Minimization , 2008, IEEE Transactions on Information Theory.

[29]  Hiroshi Ishikawa,et al.  Globally and locally consistent image completion , 2017, ACM Trans. Graph..

[30]  P. K. Bhattacharya,et al.  Kernel and Nearest-Neighbor Estimation of a Conditional Quantile , 1990 .

[31]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[32]  Simon Osindero,et al.  Conditional Generative Adversarial Nets , 2014, ArXiv.

[33]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[34]  Masaaki Imaizumi,et al.  Adaptive Approximation and Estimation of Deep Neural Network to Intrinsic Dimensionality , 2019, ArXiv.

[35]  Takafumi Kanamori,et al.  Least-Squares Conditional Density Estimation , 2010, IEICE Trans. Inf. Syst..

[36]  Xiaohong Chen,et al.  The Estimation of Conditional Densities , 2001 .

[37]  Johannes Schmidt-Hieber Nonparametric regression using deep neural networks with ReLU activation function , 2020 .

[38]  Jianqing Fan,et al.  Estimation of conditional densities and sensitivity measures in nonlinear dynamical systems , 1996 .

[39]  Jianqing Fan,et al.  A crossvalidation method for estimating conditional densities , 2004 .

[40]  Peter L. Bartlett,et al.  Nearly-tight VC-dimension and Pseudodimension Bounds for Piecewise Linear Neural Networks , 2017, J. Mach. Learn. Res..

[41]  A. Keziou Dual representation of Φ-divergences and applications , 2003 .

[42]  Ann B. Lee,et al.  Nonparametric Conditional Density Estimation in a High-Dimensional Regression Setting , 2016, 1604.00540.

[43]  Aapo Hyvärinen,et al.  Nonlinear independent component analysis: Existence and uniqueness results , 1999, Neural Networks.

[44]  Strong consistency of the kernel estimators of conditional density function , 1985 .

[45]  Yuling Jiao,et al.  Deep Nonparametric Regression on Approximately Low-dimensional Manifolds , 2021 .

[46]  Tuo Zhao,et al.  Nonparametric Regression on Low-Dimensional Manifolds using Deep ReLU Networks , 2019 .

[47]  Yuling Jiao,et al.  Deep Quantile Regression: Mitigating the Curse of Dimensionality Through Composition , 2021, 2107.04907.

[48]  Ker-Chau Li,et al.  Sliced Inverse Regression for Dimension Reduction , 1991 .

[49]  Honglak Lee,et al.  An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.