Proceedings of the Twenty-Eighth Annual ACM-SIAM Symposium on Discrete Algorithms 2017
DOI: 10.1137/1.9781611974782.4
|View full text |Cite
|
Sign up to set email alerts
|

Optimal Hashing-based Time-Space Trade-offs for Approximate Near Neighbors

Abstract: We show tight upper and lower bounds for time-space trade-offs for the c-approximate Near Neighbor Search problem. For the d-dimensional Euclidean space and npoint datasets, we develop a data structure with space n 1+ρu+o(1) + O(dn) and query time n ρq+o(1) + dn o(1) for every ρ u , ρ q ≥ 0 with:In particular, for the approximation c = 2 we get:• Space n 1.77... and query time n o(1) , significantly improving upon known data structures that support very fast queries [IM98, KOR00];• Space n 1.14... and query ti… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
184
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
3
2

Relationship

2
8

Authors

Journals

citations
Cited by 80 publications
(190 citation statements)
references
References 29 publications
(94 reference statements)
6
184
0
Order By: Relevance
“…We show that the latter is at least as hard as the (r, c)-ANNS with n 1 µ points and c O( log(n) log(1/ε) ) under the Hamming distance. Combined with the results of Panigrahy, Talwar, Wieder [32] and Andoni et al [5], we get non-trivial lower bounds in the cell-probe model with a single probe that captures an interesting class of algorithms based on adaptive coresets.…”
Section: Lower Bounds For Kde Problemsupporting
confidence: 65%
“…We show that the latter is at least as hard as the (r, c)-ANNS with n 1 µ points and c O( log(n) log(1/ε) ) under the Hamming distance. Combined with the results of Panigrahy, Talwar, Wieder [32] and Andoni et al [5], we get non-trivial lower bounds in the cell-probe model with a single probe that captures an interesting class of algorithms based on adaptive coresets.…”
Section: Lower Bounds For Kde Problemsupporting
confidence: 65%
“…Depending on the feature space and distance function chosen or learned by the practitioner, different fast approximate nearest neighbor search algorithms are available. These search algorithms, both for general high-dimensional feature spaces (e.g., Gionis et al 1999;Datar et al 2004;Bawa et al 2005;Andoni and Indyk 2008;Ailon and Chazelle 2009;Muja and Lowe 2009;Boytsov and Naidan 2013;Dasgupta and Sinha 2015;Mathy et al 2015;Andoni et al 2017) and specialized to image patches (e.g., Barnes et al 2009;Ta et al 2014), can rapidly determine which data points are close to each other while parallelizing across search queries. These methods often use locality-sensitive hashing (Indyk and Motwani, 1998), which comes with a theoretical guarantee on approximation accuracy, or randomized trees (e.g., Bawa et al 2005;Muja and Lowe 2009;Dasgupta and Sinha 2015;Mathy et al 2015), which quickly prune search spaces when the trees are sufficiently balanced.…”
Section: Explaining the Popularity Of Nearest Neighbor Methodsmentioning
confidence: 99%
“…One such application is given in high-dimensional settings where the exact nearest neighbor search is computationally expensive and Approximate Nearest Neighbor (ANN) search is often replaced in order to reduce this cost. Our flexible result allows us to use the state-of-the-art ANN algorithms (see e.g., Andoni et al [2017Andoni et al [ , 2018) while maintaining consistency and asymptotic normality.…”
Section: Adaptive Choice For Smentioning
confidence: 99%