Space and Time Efficient Kernel Density Estimation in High Dimensions

Recently, Charikar and Siminelakis (2017) presented a framework for kernel density estimation in provably sublinear query time, for kernels that possess a certain hashing-based property. However, their data structure requires a significantly increased super-linear storage space, as well as super-linear preprocessing time. These limitations inhibit the practical applicability of their approach on large datasets. In this work, we present an improvement to their framework that retains the same query time, while requiring only linear space and linear preprocessing time. We instantiate our framework with the Laplacian and Exponential kernels, two popular kernels which possess the aforementioned property. Our experiments on various datasets verify that our approach attains accuracy and query time similar to Charikar and Siminelakis (2017), with significantly improved space and preprocessing time.

[1]  Leslie Greengard,et al.  The Fast Gauss Transform , 1991, SIAM J. Sci. Comput..

[2]  Piotr Indyk,et al.  Approximate nearest neighbors: towards removing the curse of dimensionality , 1998, STOC '98.

[3]  Denis J. Dean,et al.  Comparative accuracies of artificial neural networks and discriminant analysis in predicting forest cover types from cartographic variables , 1999 .

[4]  David Haussler,et al.  Using the Fisher Kernel Method to Detect Remote Protein Homologies , 1999, ISMB.

[5]  Andrew W. Moore,et al.  'N-Body' Problems in Statistical Learning , 2000, NIPS.

[6]  Nicole Immorlica,et al.  Locality-sensitive hashing scheme based on p-stable distributions , 2004, SCG '04.

[7]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[8]  Alexandr Andoni,et al.  Near-Optimal Hashing Algorithms for Approximate Nearest Neighbor in High Dimensions , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).

[9]  Benjamin Recht,et al.  Random Features for Large-Scale Kernel Machines , 2007, NIPS.

[10]  Alexander J. Smola,et al.  Super-Samples from Kernel Herding , 2010, UAI.

[11]  Ravi Kumar,et al.  LSH-Preserving Functions and Their Applications , 2012, SODA.

[12]  Feifei Li,et al.  Quality and efficiency for kernel density estimates in large data , 2013, SIGMOD '13.

[13]  Jeff M. Phillips,et al.  Є-Samples for Kernels , 2013, SODA.

[14]  Jeffrey Pennington,et al.  GloVe: Global Vectors for Word Representation , 2014, EMNLP.

[15]  William B. March,et al.  ASKIT: Approximate Skeletonization Kernel-Independent Treecode in High Dimensions , 2014, SIAM J. Sci. Comput..

[16]  Moses Charikar,et al.  Hashing-Based-Estimators for Kernel Density in High Dimensions , 2017, 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS).

[17]  Peter Bailis,et al.  Scalable Kernel Density Classification via Threshold-Based Pruning , 2017, SIGMOD Conference.

[18]  Anshumali Shrivastava,et al.  Lsh-Sampling breaks the Computational chicken-and-egg Loop in adaptive stochastic Gradient estimation , 2018, ICLR.

[19]  Chen Luo,et al.  Arrays of (locality-sensitive) Count Estimators (ACE): Anomaly Detection on the Edge , 2018, WWW.

[20]  Jeff M. Phillips,et al.  Near-Optimal Coresets of Kernel Density Estimates , 2018, Discrete & Computational Geometry.

[21]  Chen Luo,et al.  Scaling-up Split-Merge MCMC with Locality Sensitive Sampling (LSS) , 2018, AAAI.

[22]  Philip Levis,et al.  Rehashing Kernel Evaluation in High Dimensions , 2019, ICML.