2017
DOI: 10.1145/3083897
|View full text |Cite
|
Sign up to set email alerts
|

On the Hardness and Approximation of Euclidean DBSCAN

Abstract: DBSCAN is a method proposed in 1996 for clustering multi-dimensional points, and has received extensive applications. Its computational hardness is still unsolved to this date. The original KDD‚96 paper claimed an algorithm of O ( n log n ) ”average runtime complexity„ (where n is the number of data points) without a rigorous proof. In 2013, a genuine O ( n log … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
43
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 44 publications
(44 citation statements)
references
References 29 publications
0
43
0
Order By: Relevance
“…One of the most widespread agglomeration methods, i.e. when two clusters are similar enough to be combined, is the Ward's method, while the Euclidean distance is the most common, standard distance measure for continuous data (Gan & Tao, 2017). In conclusion, the agglomeration allowed not only distinguishing the categories of business models and their individual types but also tracing the evolution of business models in the analysed sector.…”
Section: Methodsmentioning
confidence: 98%
“…One of the most widespread agglomeration methods, i.e. when two clusters are similar enough to be combined, is the Ward's method, while the Euclidean distance is the most common, standard distance measure for continuous data (Gan & Tao, 2017). In conclusion, the agglomeration allowed not only distinguishing the categories of business models and their individual types but also tracing the evolution of business models in the analysed sector.…”
Section: Methodsmentioning
confidence: 98%
“…If a non-empty cell contains at least MinPts points, the cell is called a core cell; moreover, because the maximum distance in the cell is ε, all points in the cell are core points, so it is not necessary to calculate the density of each point in the core cell. Based on fast DBSCAN algorithm, Gan and Tao proposed ρ-approximate DBSCAN [23] algorithm. The algorithm achieves an excellent complexity of O(n) in low dimension.…”
Section: Related Workmentioning
confidence: 99%
“…The runtime complexity of a single range query when using a sequential scan is O ( N ), resulting in a total runtime complexity of O(N2+false∑iri) in the worst case. In many practical applications, however, by using suitable index structures such as R*‐trees, range queries can be evaluated much faster than by using a sequential scan (Gan & Tao, ; Schubert, Sander, Ester, Kriegel, & Xu, ).…”
Section: Classic Algorithms For Flat Density‐based Clusteringmentioning
confidence: 99%
“…The runtime complexity of a single range query when using a sequential scan is O(N), resulting in a total runtime complexity of OðN 2 + P i r i Þ in the worst case. In many practical applications, however, by using suitable index structures such as R*-trees, range queries can be evaluated much faster than by using a sequential scan (Gan & Tao, 2017;Schubert, Sander, Ester, Kriegel, & Xu, 2017). GDBSCAN is an algorithmic framework that generalizes the notion of density-based clusters to the concept of density-connected decomposition for any type of data.…”
Section: Classic Algorithms For Flat Density-based Clusteringmentioning
confidence: 99%
See 1 more Smart Citation