2021
DOI: 10.48550/arxiv.2107.02565
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Prioritized training on points that are learnable, worth learning, and not yet learned (workshop version)

Abstract: We introduce Goldilocks Selection, a technique for faster model training which selects a sequence of training points that are "just right". We propose an information-theoretic acquisition functionthe reducible validation loss-and compute it with a small proxy model-GoldiProx-to efficiently choose training points that maximize information about the labels of a validation set. We show that the "hard" (e.g. high loss) points usually selected in the optimization literature are typically noisy, while the "easy" (e.… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 13 publications
(20 reference statements)
0
1
0
Order By: Relevance
“…It is well known that batch samples have a significant impact on the final performance of deep learning models [6], [19], [20], [28]. In recent years, there has been a surge of research in developing sample selection methods in various learning tasks, e.g.…”
Section: Introductionmentioning
confidence: 99%
“…It is well known that batch samples have a significant impact on the final performance of deep learning models [6], [19], [20], [28]. In recent years, there has been a surge of research in developing sample selection methods in various learning tasks, e.g.…”
Section: Introductionmentioning
confidence: 99%