Proceedings of the Twenty-Ninth Annual ACM-SIAM Symposium on Discrete Algorithms 2018
DOI: 10.1137/1.9781611975031.173
|View full text |Cite
|
Sign up to set email alerts
|

Improved Coresets for Kernel Density Estimates

Abstract: We study the construction of coresets for kernel density estimates. That is we show how to approximate the kernel density estimate described by a large point set with another kernel density estimate with a much smaller point set. For characteristic kernels (including Gaussian and Laplace kernels), our approximation preserves the L ∞ error between kernel density estimates within error ε, with coreset size 4/ε 2 , but no other aspects of the data, including the dimension, the diameter of the point set, or the ba… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
9
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 20 publications
(10 citation statements)
references
References 19 publications
1
9
0
Order By: Relevance
“…Gaussian, Laplacian). Furthermore, corresponding lower bounds Ω( √ d/ε) and Ω(1/ε 2 ) were proved for d 1/ε 2 [57] and d 1/ε 2 [56] respectively.…”
Section: Core-setsmentioning
confidence: 95%
See 2 more Smart Citations
“…Gaussian, Laplacian). Furthermore, corresponding lower bounds Ω( √ d/ε) and Ω(1/ε 2 ) were proved for d 1/ε 2 [57] and d 1/ε 2 [56] respectively.…”
Section: Core-setsmentioning
confidence: 95%
“…The literature has mostly been focused on obtaining additive error ε > 0. A general upper bound of O( 1 ε 2 ) was shown [16,56] on coresets for characteristic kernels (e.g. Gaussian, Laplacian) using a greedy construction (kernel herding [25]).…”
Section: Core-setsmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, Backurs et al [2019] show that for the Laplace and Exponential kernels with bandwidth h = 1, e.g., the value f (y) can be computed with multiplicative 1 ± ε error in time O( d √ τ ε 2 ) even in worst case over the dataset, where τ is a uniform lower bound on the KDE. Another effective approach to this problem in high dimensions is through coresets [Agarwal et al, 2005, Clarkson, 2010, Phillips and Tai, 2018a…”
Section: Background and Related Workmentioning
confidence: 99%
“…Coresets were originally proposed in computational geometry that provide strong approximation guarantees. There has been extensive work on coresets for various ML models such as K-Means [8,25,18,48], GMM [17], kernel density estimation [46], logistic regression [28,24], SVM [10,52,51], Bayesian networks [42] and so on. A recent work [56] also used the concept for coresets.…”
Section: Related Workmentioning
confidence: 99%