Proceedings of the 20th International Conference on World Wide Web 2011
DOI: 10.1145/1963405.1963487
|View full text |Cite
|
Sign up to set email alerts
|

Efficient k-nearest neighbor graph construction for generic similarity measures

Abstract: K-Nearest Neighbor Graph (K-NNG) construction is an important operation with many web related applications, including collaborative filtering, similarity search, and many others in data mining and machine learning. Existing methods for K-NNG construction either do not scale, or are specific to certain similarity measures. We present NN-Descent, a simple yet efficient algorithm for approximate K-NNG construction with arbitrary similarity measures. Our method is based on local search, has minimal space overhead … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
477
0
1

Year Published

2012
2012
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 462 publications
(481 citation statements)
references
References 24 publications
3
477
0
1
Order By: Relevance
“…Our Constant-Size Least Popular sampling policy (LP for short) can be applied to any KNN graph construction algorithm [4,5,10]. For simplicity, we apply it to a brute force approach that compares each pair of users and keeps the k most similar for each user.…”
Section: Baseline Algorithms and Competitorsmentioning
confidence: 99%
See 1 more Smart Citation
“…Our Constant-Size Least Popular sampling policy (LP for short) can be applied to any KNN graph construction algorithm [4,5,10]. For simplicity, we apply it to a brute force approach that compares each pair of users and keeps the k most similar for each user.…”
Section: Baseline Algorithms and Competitorsmentioning
confidence: 99%
“…For applications for which data freshness is more valuable than the exactness of the results, such as news recommenders, such computation time is prohibitive. To overcome these costs, most applications therefore compute an approximate KNN graph by using preindexing mechanisms [5,11] or by exploiting greedy incremental strategies [4,10] to reduce the number of similarity computations. However, it seems hard to lower even further that number.…”
Section: Introductionmentioning
confidence: 99%
“…We use the number of candidates and the number of full similarity computations as an architecture-and programming language-independent way to measure similarity search cost [33,45,46]. A naïve method may compute up to n(n − 1) = O(n 2 ) similarities to solve the APSS problem.…”
Section: Performance Measuresmentioning
confidence: 99%
“…Then we use p -nearest neighbor method to convert the similarity matrix to p-nearest neighbor graph [10]. In the N N  similarity matrix, ( , ) w i j represents the similarity between the image i and image j .…”
Section: Graph Constructionmentioning
confidence: 99%