2016 IEEE International Conference on Big Data (Big Data) 2016
DOI: 10.1109/bigdata.2016.7840682
|View full text |Cite
|
Sign up to set email alerts
|

Fast nearest neighbor search through sparse random projections and voting

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
30
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
3
2
1

Relationship

2
4

Authors

Journals

citations
Cited by 34 publications
(31 citation statements)
references
References 21 publications
1
30
0
Order By: Relevance
“…Figure 2 shows the accuracy-speed trade-off for all combinations of the considered tree types and search methods on two benchmark data sets. For RP trees, the results are in line with previous experiments [6]. For each type of tree, voting outperforms priority queue (for a given recall level, its query time is faster).…”
Section: Voting Searchsupporting
confidence: 87%
See 2 more Smart Citations
“…Figure 2 shows the accuracy-speed trade-off for all combinations of the considered tree types and search methods on two benchmark data sets. For RP trees, the results are in line with previous experiments [6]. For each type of tree, voting outperforms priority queue (for a given recall level, its query time is faster).…”
Section: Voting Searchsupporting
confidence: 87%
“…Our approach is based on exploiting the structure of randomized spacepartitioning trees [14,13,3,6]. ANN algorithms based on randomized space-partitioning trees have been used recently for example in machine translation [5], object detection [1] and recommendation engines [17].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Typically, the sparsity parameter a can be chosen as 1 √ d , as in [3], to obtain good accuracy. Now, each d-dimensional data-point p ∈ X is then projected onto the sparse vector r. The dataset X is then divided into two subsets at the median point of the projected values.…”
Section: A Mrpt Algorithmmentioning
confidence: 99%
“…We consider the problem of finding the k nearest neighbors of a query point in a given highdimensional dataset. To solve this problem efficiently, our goal is to speed up an existing algorithm [3] by parallelizing it, and to make it resilient to stragglers [4]. The k-nearest neighbor (k-NN) problem is often a first step used in a variety of real world applications including genomics [5], personalized search [6], network security [7], and web based recommendation systems [8].…”
Section: Introductionmentioning
confidence: 99%