2022
DOI: 10.48550/arxiv.2205.10733
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

GraB: Finding Provably Better Data Permutations than Random Reshuffling

Abstract: Random reshuffling, which randomly permutes the dataset each epoch, is widely adopted in model training because it yields faster convergence than with-replacement sampling. Recent studies indicate greedily chosen data orderings can further speed up convergence empirically, at the cost of using more computation and memory. However, greedy ordering lacks theoretical justification and has limited utility due to its non-trivial memory and computation overhead. In this paper, we first formulate an example-ordering … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 24 publications
(32 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?