2019
DOI: 10.48550/arxiv.1910.01215
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ES-MAML: Simple Hessian-Free Meta Learning

Abstract: We introduce ES-MAML, a new framework for solving the model agnostic meta learning (MAML) problem based on Evolution Strategies (ES). Existing algorithms for MAML are based on policy gradients, and incur significant difficulties when attempting to estimate second derivatives using backpropagation on stochastic policies. We show how ES can be applied to MAML to obtain an algorithm which avoids the problem of estimating second derivatives, and is also conceptually simple and easy to implement. Moreover, ES-MAML … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
23
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(23 citation statements)
references
References 11 publications
0
23
0
Order By: Relevance
“…In [5], the authors noted that the second derivatives can be omitted, reducing MAML to first-order MAML (FOMAML). To address the problem of high computational cost, several other first-order approximations of MAML are proposed, including Reptile in [7], Hessian-free MAML (HF-MAML) in [8] and Evolution-Strategies MAML (ES-MAML) in [9]. In our study, FOMAML is considered in the proposed framework for reducing the computational cost.…”
Section: B Related Workmentioning
confidence: 99%
“…In [5], the authors noted that the second derivatives can be omitted, reducing MAML to first-order MAML (FOMAML). To address the problem of high computational cost, several other first-order approximations of MAML are proposed, including Reptile in [7], Hessian-free MAML (HF-MAML) in [8] and Evolution-Strategies MAML (ES-MAML) in [9]. In our study, FOMAML is considered in the proposed framework for reducing the computational cost.…”
Section: B Related Workmentioning
confidence: 99%
“…Meta-learning A surge of recent works have been devoted to developing theory and algorithms of MAML [2,5,13,14,17]. For example, the 'Almost No Inner Loop' (ANIL) algorithm was proposed in [5], which dissects the meta-learning into two phases: training the initialization of a meta-model, and partial fine-tuning the classification head of the meta-model.…”
Section: Related Workmentioning
confidence: 99%
“…In supervised learning, many approaches to few-shot learning have been developed, ranging from using simple hand-crafted label signatures as priors [30] to more complex metric-learning [31] or meta-learning based [32] methods. In Reinforcement Learning, few-shot learning is almost always approached via meta-learning [33], [34], [35], [36], [37], [38], [39]. Earlier definitions of meta-learning include any algorithm that changes the meta-parameters of another learning algorithm, such that the performance of the latter is improved [40].…”
Section: Related Workmentioning
confidence: 99%
“…This not only makes the method more suitable for deceptive or sparse reward problems, but also leaves more freedom in the design and addition of additional meta-objectives. Second, like [33], [37], it is agnostic to the underlying models that are being trained. Finally, it is simple to implement and scale.…”
Section: Related Workmentioning
confidence: 99%