2022
DOI: 10.48550/arxiv.2212.04858
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Predictor networks and stop-grads provide implicit variance regularization in BYOL/SimSiam

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…Locally instructive cues may arise from the supervisory action of other brain areas, as assumed in computational models of error-driven learning 55,56,57 . Alternatively, locally instructive cues may be features of the late-stage dynamics of response, as assumed in models of self-supervised learning 58 . Future work will be required to formally address these possibilities, and to gain a mechanistic understanding of the behavioral manifestations of BTSP in the PFC.…”
Section: Discussionmentioning
confidence: 99%
“…Locally instructive cues may arise from the supervisory action of other brain areas, as assumed in computational models of error-driven learning 55,56,57 . Alternatively, locally instructive cues may be features of the late-stage dynamics of response, as assumed in models of self-supervised learning 58 . Future work will be required to formally address these possibilities, and to gain a mechanistic understanding of the behavioral manifestations of BTSP in the PFC.…”
Section: Discussionmentioning
confidence: 99%
“…Its structure is similar to BYOL, retaining the predictor of the online network, but without EMA (exponential moving average). Simsiam proves that EMA is not necessary to prevent collapse but removing it will sacrifice part of the accuracy [32].…”
Section: Contrastive Learningmentioning
confidence: 99%