2022
DOI: 10.48550/arxiv.2206.10139
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Insights into Pre-training via Simpler Synthetic Tasks

Abstract: Pre-training produces representations that are effective for a wide range of downstream tasks, but it is still unclear what properties of pre-training are necessary for effective gains. Notably, recent work shows that even pre-training on synthetic tasks can achieve significant gains in downstream tasks. In this work, we perform three experiments that iteratively simplify pre-training and show that the simplifications still retain much of its gains. First, building on prior work, we perform a systematic evalua… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 13 publications
0
1
0
Order By: Relevance
“…47 We also note that Huang et al have used atomic energies to train NN-based atomistic models. 69 In a wider perspective, the pre-training of NN models is a well-documented approach in the ML literature for various applications and domains, [70][71][72][73][74] and it has very recently been described in the context of interatomic potential models, 47,75,76 property prediction with synthetic pre-training data, 77 and as a means to learn generalpurpose representations for atomistic structure. 76…”
Section: Digital Discovery Accepted Manuscriptmentioning
confidence: 99%
“…47 We also note that Huang et al have used atomic energies to train NN-based atomistic models. 69 In a wider perspective, the pre-training of NN models is a well-documented approach in the ML literature for various applications and domains, [70][71][72][73][74] and it has very recently been described in the context of interatomic potential models, 47,75,76 property prediction with synthetic pre-training data, 77 and as a means to learn generalpurpose representations for atomistic structure. 76…”
Section: Digital Discovery Accepted Manuscriptmentioning
confidence: 99%