2016
DOI: 10.1109/tkde.2016.2539943
|View full text |Cite
|
Sign up to set email alerts
|

Improved Practical Matrix Sketching with Guarantees

Abstract: Matrices have become essential data representations for many large-scale problems in data analytics, and hence matrix sketching is a critical task. Although much research has focused on improving the error/size tradeoff under various sketching paradigms, the many forms of error bounds make these approaches hard to compare in theory and in practice. This paper attempts to categorize and compare most known methods under row-wise streaming updates with provable guarantees, and then to tweak some of these methods … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
20
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 18 publications
(21 citation statements)
references
References 63 publications
1
20
0
Order By: Relevance
“…Fiber subset selection, also called tensor cross approximation (TCA), finds a small subset of fibers which approximates the entire data tensor. For the matrix case, this problem is known as the Column/Row Subset Selection or CUR Problem which has been thoroughly investigated and for which there exist several algorithms with almost matching lower bounds [64,82,140].…”
Section: Tensor Sketching Using Tucker Modelmentioning
confidence: 99%
“…Fiber subset selection, also called tensor cross approximation (TCA), finds a small subset of fibers which approximates the entire data tensor. For the matrix case, this problem is known as the Column/Row Subset Selection or CUR Problem which has been thoroughly investigated and for which there exist several algorithms with almost matching lower bounds [64,82,140].…”
Section: Tensor Sketching Using Tucker Modelmentioning
confidence: 99%
“…Another similar line of work is the CUR factorization [4,10,12,14,27] where methods select c columns and r rows of A to form matrices C ∈ R n×c , R ∈ R r×d and U ∈ R c×r , and constructs the sketch as B = CU R. The only instance of this group that runs in input sparsity time is [4] Random projection techniques These techniques [31,36,35,26] operate data-obliviously and maintain a r×d matrix B = SA using a r×n random matrix S which has the Johnson-Lindenstrauss Transform (JLT) property [28]. Random projection methods work in the streaming model, are computationally efficient, and sufficiently accurate in practice [7]. The state-of-the-art method of this approach is by Clarkson and Woodruff [6] which was later improved slightly in [30].…”
Section: Matrix Sketching Prior Artmentioning
confidence: 99%
“…Examples of these methods include different version of iterative SVD [19,21,23,5,33]. These, however, do not have theoretical guarantees [7]. The FrequentDirections algorithm [24] is a unique in this group in that it offers strong error guarantees.…”
Section: Matrix Sketching Prior Artmentioning
confidence: 99%
“…This is achieved by using a "forgetting factor" in Step b of Algorithm 1. Such an extension is crucial, as there are pathological examples where (static) MOSES and iSVD both fail to follow the changes in the distribution of data [29]. This important research direction is left for future work.…”
Section: Prior Artmentioning
confidence: 99%