2021
DOI: 10.48550/arxiv.2102.02504
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Meta-strategy for Learning Tuning Parameters with Guarantees

Dimitri Meunier,
Pierre Alquier

Abstract: Online gradient methods, like the online gradient algorithm (OGA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows to learn the initialization and the step size in OGA with guarantees. We provide a regret analysis of the strategy in the case of convex losses. It suggests that, when there are parame… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 12 publications
0
3
0
Order By: Relevance
“…Several recent works [12,17,21] introduce algorithms for online-within-online meta-learning with provable regret bounds. Each of these approaches consists of an inner algorithm that is applied to each episode and an outer algorithm that updates hyper-parameters of the inner algorithm.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Several recent works [12,17,21] introduce algorithms for online-within-online meta-learning with provable regret bounds. Each of these approaches consists of an inner algorithm that is applied to each episode and an outer algorithm that updates hyper-parameters of the inner algorithm.…”
Section: Related Workmentioning
confidence: 99%
“…They quantify the problem difficulty as the generalized empirical variance of the per-episode best-in-hindsight points. Meunier and Alquier [21] further generalize this idea and propose a unified way to treat all the hyper-parameters. Denevi et al [12] propose to use a regularized empirical error as a performance measure instead of the regret, and their outer algorithm updates the initialization point based on the (sub-)gradient of this measure.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation