2012
DOI: 10.1007/978-3-642-32512-0_4
|View full text |Cite
|
Sign up to set email alerts
|

Improved Spectral-Norm Bounds for Clustering

Abstract: Aiming to unify known results about clustering mixtures of distributions under separation conditions, Kumar and Kannan [KK10] introduced a deterministic condition for clustering datasets. They showed that this single deterministic condition encompasses many previously studied clustering assumptions. More specifically, their proximity condition requires that in the target k-clustering, the projection of a point x onto the line joining its cluster center µ and some other center µ ′ , is a large additive factor c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

4
129
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 66 publications
(133 citation statements)
references
References 21 publications
(47 reference statements)
4
129
0
Order By: Relevance
“…The n × n edge probability matrix P = ((P ij )), given by (3), represents the population counterpart of the adjacency matrix A.…”
Section: The Stochastic Block Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…The n × n edge probability matrix P = ((P ij )), given by (3), represents the population counterpart of the adjacency matrix A.…”
Section: The Stochastic Block Modelmentioning
confidence: 99%
“…In our case, M = V τ . Recent results on cluster recovery using the K-means algorithm, as given in Kumar and Kannan [15] and Awasthi and Sheffet [3], provide conditions on X and M for the success of K-means. The following lemma is implied from Theorem 3.1 in Awasthi and Sheffet [3].…”
Section: Cluster Recovery Using K-means Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…Here β is chosen to be 4 5 α since we are not able to prove the bound for β = α but will be able to when β is slightly smaller than α. Some other constant can be used instead of 4 5 . Definition 6.2.…”
Section: Sublinear Algorithm For the Min-sum Objectivementioning
confidence: 99%
“…Kumar and Kannan [19] proposed a proximity condition which assumes that in the target clustering, most data points satisfy that they are closer to their center than to any other center by an additive factor in the order of the maximal standard variance of their clusters in any direction. Their results are improved by Awasthi and Sheffet [4], which proposed a weaker version of the proximity condition called center separation, and designed algorithms achieving stronger guarantees under this weaker condition. These notions are not directly comparable to the perturbation resilience property.…”
mentioning
confidence: 99%