2009
DOI: 10.1590/s0101-74382009000300002
|View full text |Cite
|
Sign up to set email alerts
|

A branch-and-cut SDP-based algorithm for minimum sum-of-squares clustering

Abstract: Minimum sum-of-squares clustering (MSSC) consists in partitioning a given set of n points into k clusters in order to minimize the sum of squared distances from the points to the centroid of their cluster. Recently, Peng & Xia (2005) established the equivalence between 0-1 semidefinite programming (SDP) and MSSC. In this paper, we propose a branch-and-cut algorithm for the underlying 0-1 SDP model. The algorithm obtains exact solutions for fairly large data sets with computing times comparable with those of t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
22
0

Year Published

2010
2010
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 14 publications
(22 citation statements)
references
References 28 publications
0
22
0
Order By: Relevance
“…So, we do not refer to its results in the subsequent tables. As empirically observed in [2], the performance of algorithm bb-sdp deteriorates as k decreases, in contrast with…”
Section: Results In the Planementioning
confidence: 73%
See 2 more Smart Citations
“…So, we do not refer to its results in the subsequent tables. As empirically observed in [2], the performance of algorithm bb-sdp deteriorates as k decreases, in contrast with…”
Section: Results In the Planementioning
confidence: 73%
“…The results are also compared to those of two other methods proposed in the literature, i.e., the repetitive branch-and-bound algorithm (rbba) of Brusco [9] and the best branch-and-cut SDP-based algorithm (bb-sdp) of [2]. Tables 2-7 show results for data sets in the plane.…”
Section: Results In the Planementioning
confidence: 99%
See 1 more Smart Citation
“…In such a problem, each object can be considered as a point in a n-dimensional space and each cluster can be identified by its center, called centroid, a non-observable object calculated by taking the mean of all the objects assigned to this cluster [18,32,34]. To express similarity between objects, i.e., homogeneity inside a cluster, several similarity measures have been proposed, such as a metric defined on the data set [2,7]. One of the most used (dis)similarity measures is the squared Euclidean distance [3,4,14,18,34].…”
Section: Introductionmentioning
confidence: 99%
“…Most of the sophisticated computation on clustering problems has addressed the MSS; see, e.g., the most recent exact methods (Aloise and Hansen, ; Aloise et al, ). We do not discuss these approaches in detail because (i) MSS is computationally rather different from RASS and (ii) we focus here on spectral bounds (for which there is not much computation in the MSS special case).…”
Section: Introductionmentioning
confidence: 99%