2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2017
DOI: 10.1109/icassp.2017.7953382
|View full text |Cite
|
Sign up to set email alerts
|

Compressive K-means

Abstract: The Lloyd-Max algorithm is a classical approach to perform Kmeans clustering. Unfortunately, its cost becomes prohibitive as the training dataset grows large. We propose a compressive version of K-means (CKM), that estimates cluster centers from a sketch, i.e. from a drastically compressed representation of the training dataset.We demonstrate empirically that CKM performs similarly to Lloyd-Max, for a sketch size proportional to the number of centroids times the ambient dimension, and independent of the size o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
77
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
1

Relationship

4
4

Authors

Journals

citations
Cited by 37 publications
(79 citation statements)
references
References 24 publications
(46 reference statements)
2
77
0
Order By: Relevance
“…To recover the centroids C from y, the state-of-the-art algorithm is compressed learning via orthogonal matching pursuit with replacement (CL-OMPR) [5,6]. It aims to solve arg min…”
Section: A Sketched Clusteringmentioning
confidence: 99%
“…To recover the centroids C from y, the state-of-the-art algorithm is compressed learning via orthogonal matching pursuit with replacement (CL-OMPR) [5,6]. It aims to solve arg min…”
Section: A Sketched Clusteringmentioning
confidence: 99%
“…Estimating such signals from a finite number of Fourier measurements is known as the super-resolution problem [9]. Also, the estimation of spikes from random Fourier measurements is at the core of the compressive K-means algorithm were k-means cluster centers are estimated from a compressed database [21]. In the space M of finite signed measure over R d , we aim at recovering x 0 = i=1,k a i δ t i from the measurements…”
Section: Introduction 1contextmentioning
confidence: 99%
“…The sketch is thus the pooling (average) of random projections of the data samples after passing through a nonlinear, periodic signature-the complex exponential. The Compressive K-Means (CKM) method [8] clusters X from z X by replacing (1) with a sketch matching optimization problem:…”
Section: Introductionmentioning
confidence: 99%
“…For the related kernel K-means problem, [20] uses Random Fourier Features [21], i.e., the low-dimensional mapping z (·) defined in (2). For u, v ∈ R n , the inner product z u , z v approximates a shift-invariant kernel κ(u, v) associated with the frequency distribution Λ. CKM [8] actually averages individual RFF of data points. Interestingly, κ also defines a Reproducing Kernel Hilbert Space in which two probability density functions (pdfs) can be compared with a Maximum Mean Discrepancy (MMD) metric [22][23][24][25][26].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation