2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412010
|View full text |Cite
|
Sign up to set email alerts
|

Meta Learning via Learned Loss

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
37
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(38 citation statements)
references
References 7 publications
1
37
0
Order By: Relevance
“…Gradient based methods are closely related to meta-gradient methods, as they use second-order derivatives to optimize ζ. Bechtle et al (2020) described a general framework for meta learning with learned loss functions. Most meta-gradient RL algorithms, as mentioned in Section 4.5, start with a human designed loss function L(•) and modify it with parameterization L(θ; ζ) to allow inner loop and outer loop procedures.…”
Section: Learning Reinforcement Learning Algorithmsmentioning
confidence: 99%
“…Gradient based methods are closely related to meta-gradient methods, as they use second-order derivatives to optimize ζ. Bechtle et al (2020) described a general framework for meta learning with learned loss functions. Most meta-gradient RL algorithms, as mentioned in Section 4.5, start with a human designed loss function L(•) and modify it with parameterization L(θ; ζ) to allow inner loop and outer loop procedures.…”
Section: Learning Reinforcement Learning Algorithmsmentioning
confidence: 99%
“…The research of meta learning, also known as "learning to learn", focuses on training models that gain experience and improve performance over multiple learning episodes/tasks [30], which has seen wide adoption in research areas such as reinforcement learning [31,32]. This can be applied in various aspects of learning problems under a bi-level optimization framework [33], including the data set generation [34], learning objective [31,35,36], model architecture [37], initialization parameters [38], and the learning rules [32, 39? ]. Our work studies the meta learning in the context of large scale segment prediction tasks, and further investigates meta learning's application to single task learning with novel auxiliary task family construction.…”
Section: Meta Learningmentioning
confidence: 99%
“…Consider a distribution over tasks p : T → [0, 1], we assume a set of 𝑀 source training (i.e. the support in meta learning literature [36]) and validation (query in meta learning literature [36]) data-sets available sampled from T,…”
Section: Problem Overviewmentioning
confidence: 99%
“…Metalearning, aka learning to learn, and AutoML have been applied for a wide variety of purposes as summarised in [17,21]. Of particular relevance is meta-learning of loss functions, which has been studied for various purposes including providing differentiable surrogates of nondifferentiable objectives [19], optimising efficiency and asymptotic performance of learning [22,4,18,48,11,12], and improving robustness to train/test domain-shift [3,30]. We are interested in learning white-box losses -i.e., those that can be expressed a short human-readable parametric equation -for efficiency and improved task-transferability compared to neural network alternatives [4,18,3,30], which tend to be less interpretable and need to be learned taskspecifically.…”
Section: Meta-learning Automl and Loss Learningmentioning
confidence: 99%
“…Of particular relevance is meta-learning of loss functions, which has been studied for various purposes including providing differentiable surrogates of nondifferentiable objectives [19], optimising efficiency and asymptotic performance of learning [22,4,18,48,11,12], and improving robustness to train/test domain-shift [3,30]. We are interested in learning white-box losses -i.e., those that can be expressed a short human-readable parametric equation -for efficiency and improved task-transferability compared to neural network alternatives [4,18,3,30], which tend to be less interpretable and need to be learned taskspecifically. Meta-learning of white-box model components has been demonstrated for optimisers [47], activation functions [35], neural architectures [43] and losses for accelerating conventional supervised learning [11,12].…”
Section: Meta-learning Automl and Loss Learningmentioning
confidence: 99%