2022
DOI: 10.1007/s10994-022-06210-y
|View full text |Cite
|
Sign up to set email alerts
|

Stateless neural meta-learning using second-order gradients

Abstract: Meta-learning can be used to learn a good prior that facilitates quick learning; two popular approaches are MAML and the meta-learner LSTM. These two methods represent important and different approaches in meta-learning. In this work, we study the two and formally show that the meta-learner LSTM subsumes MAML, although MAML, which is in this sense less general, outperforms the other. We suggest the reason for this surprising performance gap is related to second-order gradients. We construct a new algorithm (na… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 19 publications
(22 reference statements)
0
2
0
Order By: Relevance
“…These methods aim to metalearn good settings for various hyperparameters, such as the initialization parameters, such that new tasks can be learned quickly using optimization methods. These methods vary from regular stochastic gradient descent, as used in MAML (Finn et al, 2017) and Reptile (Nichol et al, 2018), to meta-learned procedures where a network updates the weights of a base-learner (Ravi et al, 2017;Andrychowicz et al, 2016;Li et al, 2017;Rusu et al, 2019;Li & Malik, 2018;Huisman et al, 2022). SAP aims to learn good initialization parameters such that new tasks can be learned quickly with regular gradient descent, similar to MAML.…”
Section: Optimization-based Meta-learningmentioning
confidence: 99%
See 1 more Smart Citation
“…These methods aim to metalearn good settings for various hyperparameters, such as the initialization parameters, such that new tasks can be learned quickly using optimization methods. These methods vary from regular stochastic gradient descent, as used in MAML (Finn et al, 2017) and Reptile (Nichol et al, 2018), to meta-learned procedures where a network updates the weights of a base-learner (Ravi et al, 2017;Andrychowicz et al, 2016;Li et al, 2017;Rusu et al, 2019;Li & Malik, 2018;Huisman et al, 2022). SAP aims to learn good initialization parameters such that new tasks can be learned quickly with regular gradient descent, similar to MAML.…”
Section: Optimization-based Meta-learningmentioning
confidence: 99%
“…Gradient-based meta-learning methods struggle to scale well to deep networks as recent work suggests that simple pre-training and fine-tuning of the output layer (Tian et al, 2020;Chen et al, 2021;Huisman et al, 2021a) can yield superior performance on common fewshot image classification benchmarks. This is also the reason, besides searching for energyefficient few-shot learners, that in our experiments we focus on relatively shallow backbones that adapt all layers when learning new tasks, instead of only the output layer.…”
Section: Future Workmentioning
confidence: 99%