Proceedings of the 31st ACM International Conference on Information &Amp; Knowledge Management 2022
DOI: 10.1145/3511808.3557098
|View full text |Cite
|
Sign up to set email alerts
|

Approximate Nearest Neighbor Search under Neural Similarity Metric for Large-Scale Recommendation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…In some cases, we may need much more results, such as thousands. For example, in recommendation systems, a large number of candidates are first recalled and then filtered to get the final recommendations [17,67]. Fig.…”
Section: Large-scale Search Resultsmentioning
confidence: 99%
“…In some cases, we may need much more results, such as thousands. For example, in recommendation systems, a large number of candidates are first recalled and then filtered to get the final recommendations [17,67]. Fig.…”
Section: Large-scale Search Resultsmentioning
confidence: 99%
“…Moreover, work has been done to leverage current hardware capabilities, aside from algorithmic improvements. For example, when dealing with data that cannot be accommodated in memory, ANN methods like DiskANN [22] or SPANN [23] propose to use data locality and fast disk storage (Solid State Disks (SSD)). Also, multi-threading has been exploited in ANN, with examples such as SCANN [10], or through threading parallelism at the level of query processing.…”
Section: Related Workmentioning
confidence: 99%
“…For example, (Jégou et al, 2011b) and (Baranchuk et al, 2018) add another refinement stage over the quantized embeddings and skip less promising clusters according to tailored heuristics. (Chen et al, 2021) create duplicated reference for boundary embeddings to improve recall with high efficiency. The other research thread optimizes the VQ index towards retrieval quality with cross-entropy loss instead of minimizing the reconstruction loss.…”
Section: Related Workmentioning
confidence: 99%
“…By increasing the number of clusters to scan, one may expect higher retrieval quality since the relevant document is more likely to be included, yet with higher query latency since there are more documents to evaluate (Jégou et al, 2011a). On top of the basic idea, recent studies improve the accuracy of IVF by grouping the cluster embeddings and skipping the least promising groups (Baranchuk et al, 2018), creating duplicated records for boundary embeddings (Chen et al, 2021), and end-to-end learning the cluster assignments by knowledge distillation (Xiao et al, 2022a). Despite their improvements, IVF still exhibits limited retrieval quality, especially when high efficiency is needed.…”
Section: Introductionmentioning
confidence: 99%