2021
DOI: 10.48550/arxiv.2101.09536
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Memory-Efficient Semi-Supervised Continual Learning: The World is its Own Replay Buffer

Abstract: Rehearsal is a critical component for class-incremental continual learning, yet it requires a substantial memory budget. Our work investigates whether we can significantly reduce this memory budget by leveraging unlabeled data from an agent's environment in a realistic and challenging continual learning paradigm. Specifically, we explore and formalize a novel semi-supervised continual learning (SSCL) setting, where labeled data is scarce yet non-i.i.d. unlabeled data from the agent's environment is plentiful. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
8
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
7

Relationship

3
4

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 38 publications
0
8
0
Order By: Relevance
“…Avoiding privacy concerns, this work also follows a line of work that doesn't store real examples for experience replay, such as generating examples by GAN (Atkinson et al, 2018), synthesizing examples (Xu et al, 2022) by model-inversion (Smith et al, 2021b), and using unlabeled data in the learning environment (Smith et al, 2021a). In language domain, LAMOL (Sun et al, 2019) trains the language model to solve current tasks and generate current training examples simultaneously, then this model can generate "pseudo" old examples for replay before any new tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Avoiding privacy concerns, this work also follows a line of work that doesn't store real examples for experience replay, such as generating examples by GAN (Atkinson et al, 2018), synthesizing examples (Xu et al, 2022) by model-inversion (Smith et al, 2021b), and using unlabeled data in the learning environment (Smith et al, 2021a). In language domain, LAMOL (Sun et al, 2019) trains the language model to solve current tasks and generate current training examples simultaneously, then this model can generate "pseudo" old examples for replay before any new tasks.…”
Section: Related Workmentioning
confidence: 99%
“…In continual and lifelong learning there exists a sequence "tasks", each of which could be different environments, datasets, or novel classes. These tasks can be overlapping, task boundaries don't have to be welldefined, and can include a mixture of supervised and unsupervised data, but in most cases these tasks are disjoint and task boundaries are known and discrete (Parisi et al 2019;Silver, Yang, and Li 2013;Smith et al 2021). Most distinctly, in continual learning the model is trained on only one task at a time but validated on that task and all prior tasks.…”
Section: Novelty Background and Related Workmentioning
confidence: 99%
“…Another approach is to regularize the model with respect to past task knowledge while training the new task. This can either be done by regularizing the model in the weight space (i.e., penalize changes to model parameters) [1,13,30,55,63] or the prediction space (i.e., penalize changes to model predictions) [7,23,33,36,52]. Prediction space regularization (accomplished using knowledge distillation) has been found to perform better than model regularization based methods for class-incremental learning [35,57].…”
Section: Background and Related Workmentioning
confidence: 99%