2022
DOI: 10.1609/aaai.v36i7.20674
|View full text |Cite
|
Sign up to set email alerts
|

Same State, Different Task: Continual Reinforcement Learning without Interference

Abstract: Continual Learning (CL) considers the problem of training an agent sequentially on a set of tasks while seeking to retain performance on all previous tasks. A key challenge in CL is catastrophic forgetting, which arises when performance on a previously mastered task is reduced when learning a new task. While a variety of methods exist to combat forgetting, in some cases tasks are fundamentally incompatible with each other and thus cannot be learnt by a single policy. This can occur, in reinforcement learning (… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 12 publications
(12 citation statements)
references
References 31 publications
0
9
0
Order By: Relevance
“…Changepoint Detection: As discussed previously, one core issue in addressing settings with evolving task states is being able to detect change points or boundaries between significant switches without an oracle as in (Padakandla et al, 2019;Da Silva et al, 2006;Rosman and Ramamoorthy, 2012;Hadoux, Beynier, and Weng, 2014b;Li, Gu, Zhu, and Zhang, 2019;Kessler, Parker-Holder, Ball, Zohren, and Roberts, 2022;Luo, Jiang, Yu, Zhang, and Zhang, 2022). However, these approaches generally tend to be reactive to a changing distribution rather than proactive about anticipated changes in the future.…”
Section: Context Detectionmentioning
confidence: 99%
“…Changepoint Detection: As discussed previously, one core issue in addressing settings with evolving task states is being able to detect change points or boundaries between significant switches without an oracle as in (Padakandla et al, 2019;Da Silva et al, 2006;Rosman and Ramamoorthy, 2012;Hadoux, Beynier, and Weng, 2014b;Li, Gu, Zhu, and Zhang, 2019;Kessler, Parker-Holder, Ball, Zohren, and Roberts, 2022;Luo, Jiang, Yu, Zhang, and Zhang, 2022). However, these approaches generally tend to be reactive to a changing distribution rather than proactive about anticipated changes in the future.…”
Section: Context Detectionmentioning
confidence: 99%
“…We followed Kessler et al (2022) in evaluating the efficacy of a continual learning method in forgetting, measuring stability and backward transfer, and forward transfer, also acting as a measure for plasticity. These measurements were recorded for each environment in the experiment suite T = (τ 1 , τ 2 , .…”
Section: Continual Learning Assessmentmentioning
confidence: 99%
“…In the sequential task setting (Moskovitz et al, 2022a;Pacchiano et al, 2022), tasks (MDPs) are sampled one at a time from P M , with the agent training on each until convergence. In contrast to continual learning (Kessler et al, 2021), the agent's goal is simply to learn a new policy for each task more quickly as more are sampled, rather than learning a single policy which maintains its performance across tasks. Another important setting is meta-RL, which we do not consider here.…”
Section: Multiple Tasksmentioning
confidence: 99%