Uniform Convergence Rate of the Kernel Density Estimator Adaptive to Intrinsic Volume Dimension

We derive concentration inequalities for the supremum norm of the difference between a kernel density estimator (KDE) and its point-wise expectation that hold uniformly over the selection of the bandwidth and under weaker conditions on the kernel and the data generating distribution than previously used in the literature. We first propose a novel concept, called the volume dimension, to measure the intrinsic dimension of the support of a probability distribution based on the rates of decay of the probability of vanishing Euclidean balls. Our bounds depend on the volume dimension and generalize the existing bounds derived in the literature. In particular, when the data-generating distribution has a bounded Lebesgue density or is supported on a sufficiently well-behaved lower-dimensional manifold, our bound recovers the same convergence rate depending on the intrinsic dimension of the support as ones known in the literature. At the same time, our results apply to more general cases, such as the ones of distribution with unbounded densities or supported on a mixture of manifolds with different dimensions. Analogous bounds are derived for the derivative of the KDE, of any order. Our results are generally applicable but are especially useful for problems in geometric inference and topological data analysis, including level set estimation, density-based clustering, modal clustering and mode hunting, ridge estimation and persistent homology.

[1]  E. Giné,et al.  Rates of strong uniform consistency for multivariate kernel density estimators , 2002 .

[2]  Larry A. Wasserman,et al.  Minimax Rates for Estimating the Dimension of a Manifold , 2016, J. Comput. Geom..

[3]  O. Bousquet A Bennett concentration inequality and its application to suprema of empirical processes , 2002 .

[4]  José-Manuel Rey,et al.  Dimension of a measure , 2000 .

[5]  A. Rinaldo,et al.  Generalized density clustering , 2009, 0907.3454.

[6]  Stephen Smale,et al.  Finding the Homology of Submanifolds with High Confidence from Random Samples , 2008, Discret. Comput. Geom..

[7]  Larry A. Wasserman,et al.  Risk Bounds For Mode Clustering , 2015, ArXiv.

[8]  Yen-Chi Chen,et al.  Density Level Sets: Asymptotics, Inference, and Visualization , 2015, 1504.05438.

[9]  Heinrich Jiang,et al.  Uniform Convergence Rates for Kernel Density Estimation , 2017, ICML.

[10]  Sivaraman Balakrishnan,et al.  Confidence sets for persistence diagrams , 2013, The Annals of Statistics.

[11]  Frédéric Chazal,et al.  Robust Topological Inference: Distance To a Measure and Kernel Distance , 2014, J. Mach. Learn. Res..

[12]  Sivaraman Balakrishnan,et al.  Cluster Trees on Manifolds , 2013, NIPS.

[13]  Sivaraman Balakrishnan,et al.  Statistical Inference for Cluster Trees , 2016, NIPS.

[14]  Larry A. Wasserman,et al.  Optimal Ridge Detection using Coverage Risk , 2015, NIPS.

[15]  José E. Chacón,et al.  A Population Background for Nonparametric Density-Based Clustering , 2014, 1408.1381.

[16]  D. Pollard,et al.  $U$-Processes: Rates of Convergence , 1987 .

[17]  E. Parzen On Estimation of a Probability Density Function and Mode , 1962 .

[18]  David Mason,et al.  On the Estimation of the Gradient Lines of a Density and the Consistency of the Mean-Shift Algorithm , 2016, J. Mach. Learn. Res..

[19]  Ingo Steinwart,et al.  Adaptive Clustering Using Kernel Density Estimators , 2017, J. Mach. Learn. Res..

[20]  Andreas Christmann,et al.  Support vector machines , 2008, Data Mining and Knowledge Discovery Handbook.

[21]  Ronald F. Gariepy FUNCTIONS OF BOUNDED VARIATION AND FREE DISCONTINUITY PROBLEMS (Oxford Mathematical Monographs) , 2001 .

[22]  Joel Zinn,et al.  Weighted uniform consistency of kernel density estimators , 2004 .

[23]  Armelle Guillou,et al.  Laws of the Iterated Logarithm for Censored Data , 1999 .

[24]  Prakasa Rao Nonparametric functional estimation , 1983 .

[25]  Uwe Einmahl,et al.  Uniform in bandwidth consistency of kernel-type function estimators , 2005 .

[26]  Ingo Steinwart,et al.  Consistency and Rates for Clustering with DBSCAN , 2012, AISTATS.

[27]  F. Bach,et al.  Sharp asymptotic and finite-sample rates of convergence of empirical measures in Wasserstein distance , 2017, Bernoulli.

[28]  Michel Verleysen,et al.  Nonlinear Dimensionality Reduction , 2021, Computer Vision.

[29]  E. Giné,et al.  On consistency of kernel density estimators for randomly censored data: rates holding uniformly over adaptive intervals , 2001 .

[30]  Christopher R. Genovese,et al.  Asymptotic theory for density ridges , 2014, 1406.5663.