2018
DOI: 10.48550/arxiv.1812.06080
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Reconciling meta-learning and continual learning with online mixtures of tasks

Abstract: Learning-to-learn or meta-learning leverages data-driven inductive bias to increase the efficiency of learning on a novel task. This approach encounters difficulty when transfer is not advantageous, for instance, when tasks are considerably dissimilar or change over time. We use the connection between gradient-based meta-learning and hierarchical Bayes to propose a Dirichlet process mixture of hierarchical Bayesian models over the parameters of an arbitrary parametric model such as a neural network. In contras… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 31 publications
0
3
0
Order By: Relevance
“…Javed and White [203] proposed a meta-objective to learn representations which are naturally highly sparse and thus effectively mitigate catastrophic forgetting in continual learning. Jerfel et al [204] employed a meta-learner to control the amount of knowledge transfer between tasks and automatically adapt to a new task when a task distribution shift is detected. Rajasegaran et al [205] proposed an Incremental Task-Agnostic Meta-learning (iTAML) method, which adapts meta learning to separate the task-agnostic feature extractor from the task-specific classifier.…”
Section: Other Methodsmentioning
confidence: 99%
“…Javed and White [203] proposed a meta-objective to learn representations which are naturally highly sparse and thus effectively mitigate catastrophic forgetting in continual learning. Jerfel et al [204] employed a meta-learner to control the amount of knowledge transfer between tasks and automatically adapt to a new task when a task distribution shift is detected. Rajasegaran et al [205] proposed an Incremental Task-Agnostic Meta-learning (iTAML) method, which adapts meta learning to separate the task-agnostic feature extractor from the task-specific classifier.…”
Section: Other Methodsmentioning
confidence: 99%
“…MAML aims at finding a set of joined task parameters that can be easily fine-tuned to new test tasks via few gradient descent updates. MAML can also be treated as a Bayesian hierarchical model [10,15,18]. Bayesian MAML [55] combines efficient gradient-based meta-learning with non-parametric variational inference in a principled probabilistic framework.…”
Section: Related Workmentioning
confidence: 99%
“…However, standard meta-learning algorithms operate in batch mode, making them poorly suited for continuously evolving environments. More recently, online meta-learning methods have been proposed with the goal of enabling continual adaptation Jerfel et al, 2018;Yao et al, 2020;Nagabandi et al, 2018;Li & Hospedales, 2020), where a constant stream of data from distinct tasks is used for both adaptation and meta-training. In this scheme, meta-training is used to accelerate how quickly the network can adapt to each new task it sees, and simultaneously use that data from each new task for meta-training.…”
Section: Introductionmentioning
confidence: 99%