Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data 2015
DOI: 10.1145/2723372.2747647
|View full text |Cite
|
Sign up to set email alerts
|

Lemp

Abstract: We study the problem of efficiently retrieving large entries in the product of two given matrices, which arises in a number of data mining and information retrieval tasks. We focus on the setting where the two input matrices are tall and skinny, i.e., with millions of rows and tens to hundreds of columns. In such settings, the product matrix is large and its complete computation is generally infeasible in practice. To address this problem, we propose the LEMP algorithm, which efficiently retrieves only the lar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
5
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(5 citation statements)
references
References 21 publications
0
5
0
Order By: Relevance
“…For example, the use of factor models in recommender systems leads to matrix products B T C with B, C ∈ R m×n , m n, and n very large [29]. Another application is link prediction in graphs [26].…”
Section: C585 the P Largest Entries Of Ementioning
confidence: 99%
See 1 more Smart Citation
“…For example, the use of factor models in recommender systems leads to matrix products B T C with B, C ∈ R m×n , m n, and n very large [29]. Another application is link prediction in graphs [26].…”
Section: C585 the P Largest Entries Of Ementioning
confidence: 99%
“…An important feature of our algorithms is that they can be treated as black boxes that can be applied to many different problems, in contrast to the more specialized algorithms designed for products of two matrices, such as those in [3], [29]. Since our algorithms only require the computation of matrix-vector products, they are relatively simple to implement and can serve as a benchmark for testing more specialized algorithms in multiple application areas.…”
mentioning
confidence: 99%
“…The inner product search (IPS) problem is important in many fields, e.g., information retrieval [18,32,35], recommender systems [6,7,30], data mining [19,27], databases [22,24,29], artificial intelligence [13,36], and machine learning [14,23,34]. These fields usually consider the 𝑘 maximum inner product search (𝑘-MIPS) problem, which, given a query vector and an output size 𝑘, returns the 𝑘 vectors having the maximum inner product with the query.…”
Section: Introductionmentioning
confidence: 99%
“…○ We iterate the above operation until we have |Q 𝑘 | = 𝑘 2. State-of-the-art exact IPS algorithms are based on linear scan[22,29]. Trivially, approximation algorithms which do not guarantee that all vectors such that x • q ≥ 𝜏 are included in the search result cannot solve the FI-IPS problem correctly.…”
mentioning
confidence: 99%
“…This paper considers the top-k inner product join problem, which is defined as follows: Given two sets X and Y of high-dimensional vectors and a result size k, top-k inner product join between X and Y retrieves k pairs of vectors x, y , where x ∈ X and y ∈ Y, with the largest inner product among X × Y. This problem has important applications, such as recommendation [1]- [3], information extraction [4], and finding outlier correlations [5]. More specifically, Figure 1 depicts a histogram of inner products of 1 million randomly sampled vector pairs in a Yahoo!…”
Section: Introductionmentioning
confidence: 99%