Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2016
DOI: 10.1145/2939672.2939779
|View full text |Cite
|
Sign up to set email alerts
|

Overcoming Key Weaknesses of Distance-based Neighbourhood Methods using a Data Dependent Dissimilarity Measure

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
46
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 49 publications
(46 citation statements)
references
References 19 publications
0
46
0
Order By: Relevance
“…First, mass-based dissimilarity measures 36,37 have been shown to outperform distance measures using the same NN algorithms in classification, clustering, anomaly detection, and information retrieval tasks. First, mass-based dissimilarity measures 36,37 have been shown to outperform distance measures using the same NN algorithms in classification, clustering, anomaly detection, and information retrieval tasks.…”
Section: Discussionmentioning
confidence: 99%
“…First, mass-based dissimilarity measures 36,37 have been shown to outperform distance measures using the same NN algorithms in classification, clustering, anomaly detection, and information retrieval tasks. First, mass-based dissimilarity measures 36,37 have been shown to outperform distance measures using the same NN algorithms in classification, clustering, anomaly detection, and information retrieval tasks.…”
Section: Discussionmentioning
confidence: 99%
“…Previous works [7,28] have shown that random partitions of data can be used to compute a similarity between the instances. In particular, in Unsupervised Extremely Randomized Trees (UET), the idea is that all instances ending up in the same leaves are more similar to each other than to other instances.…”
Section: Methodsmentioning
confidence: 99%
“…The intuition behind our proposed method, GT, is to leverage a similar partition in the vertices of a graph. Instead of using the similarity computation that we described previously, we chose to use the mass-based approach introduced by Ting et al [28] instead. The key property of their measure is that the dissimilarity between two instances in a dense region is higher than the same interpoint dissimilarity between two instances in a sparse region of the same space.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…They are sampled from the data distribution of each class. In consequence, the trees are still learning an abstraction of the data, using the trees as a density estimator [46].…”
Section: How To Learn a Proximity Forest?mentioning
confidence: 99%