1973
DOI: 10.1109/t-c.1973.223640
|View full text |Cite
|
Sign up to set email alerts
|

Clustering Using a Similarity Measure Based on Shared Near Neighbors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
482
0
8

Year Published

1986
1986
2015
2015

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 873 publications
(490 citation statements)
references
References 9 publications
0
482
0
8
Order By: Relevance
“…Early approaches exploited the so called guilt-by-association (GBA) rule, which makes predictions based on the majority or weighted majority of labels in the direct neighborhood, assuming that interacting nodes are likely to share similar properties [38,57]. Analogously, k-nearest neighborhood (kNN) methods consider only the labels of the k most similar neighbors [32]; in turn, shared similarity metrics, as those proposed in [28,18], can be introduced to generalize the notion of pairwise-similarity among nodes by taking into account the contribution of shared neighbors [14,9]. Other methodologies predict labels by propagating node labels to neighbors with an iterative process until convergence [70,69], or by evaluating the functional flows through the nodes of the graph [62,49].…”
Section: Introductionmentioning
confidence: 99%
“…Early approaches exploited the so called guilt-by-association (GBA) rule, which makes predictions based on the majority or weighted majority of labels in the direct neighborhood, assuming that interacting nodes are likely to share similar properties [38,57]. Analogously, k-nearest neighborhood (kNN) methods consider only the labels of the k most similar neighbors [32]; in turn, shared similarity metrics, as those proposed in [28,18], can be introduced to generalize the notion of pairwise-similarity among nodes by taking into account the contribution of shared neighbors [14,9]. Other methodologies predict labels by propagating node labels to neighbors with an iterative process until convergence [70,69], or by evaluating the functional flows through the nodes of the graph [62,49].…”
Section: Introductionmentioning
confidence: 99%
“…For computing the nearest neighbors in high dimensional data, SNN measures have been reported to be effective in practice, and supposedly less prone to the curse of dimensionality than conventional distance measures. SNN measures have found use in the design of merge criteria of agglomerative clustering algorithms [25,27,28], in approaches for clustering high-dimensional data sets [26,29], and in finding outliers in subspaces of high dimensional data [30]. However, in all of these studies, no systematic investigation has been made into the advantages of SNN measures over conventional distance measures for high-dimensional data.…”
Section: Introductionmentioning
confidence: 99%
“…[8]). A proximity measure that is better suited for multidimensional data was proposed in [6]. In this paper, proximity between a pair of points was defined to be the number of neighbors they share.…”
Section: Estimating Proximity In Multidimensional Spaces With Sparse mentioning
confidence: 99%
“…The algorithm, named Clustering With Nearest Neighborhood (CWNN), is inspired by ideas presented in [2], [6] and [7]. CWNN employs the SNN graph to detect the so-called core data points.…”
Section: Introductionmentioning
confidence: 99%