2019 IEEE/CVF International Conference on Computer Vision (ICCV) 2019
DOI: 10.1109/iccv.2019.00040
|View full text |Cite
|
Sign up to set email alerts
|

Overcoming Catastrophic Forgetting With Unlabeled Data in the Wild

Abstract: Lifelong learning with deep neural networks is wellknown to suffer from catastrophic forgetting: the performance on previous tasks drastically degrades when learning a new task. To alleviate this effect, we propose to leverage a large stream of unlabeled data easily obtainable in the wild. In particular, we design a novel classincremental learning scheme with (a) a new distillation loss, termed global distillation, (b) a learning strategy to avoid overfitting to the most recent task, and (c) a confidence-based… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
136
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 151 publications
(145 citation statements)
references
References 32 publications
1
136
0
Order By: Relevance
“…Catastrophic forgetting appears when weights in neural networks trained on previous tasks change with learning new ones. In order to resist forgetting, these methods adopt dynamic architecture (architecture grows with new tasks) [22], memorizing the representative samples of learned tasks for rehearsal [23], [24], preventing significant changes in the representation for previous tasks [25], [26], or Bayesian methods [27], [28]. These methods alleviate catastrophic forgetting in continual learning for neural networks, however improving performance through learning multiple related tasks has not been achieved.…”
Section: Related Workmentioning
confidence: 99%
“…Catastrophic forgetting appears when weights in neural networks trained on previous tasks change with learning new ones. In order to resist forgetting, these methods adopt dynamic architecture (architecture grows with new tasks) [22], memorizing the representative samples of learned tasks for rehearsal [23], [24], preventing significant changes in the representation for previous tasks [25], [26], or Bayesian methods [27], [28]. These methods alleviate catastrophic forgetting in continual learning for neural networks, however improving performance through learning multiple related tasks has not been achieved.…”
Section: Related Workmentioning
confidence: 99%
“…For experimental validations, we use CIFAR100 [48] as it is the most popular benchmark dataset for the IL algorithms (also called iCIFAR100) [3], [7], [9], [20], TinyIma-geNet [49] used in [50], and a subset of ImageNet (ILSVRC 2012) dataset [51] with 100 classes used in [7], [9], which we call as ImageNet-100.…”
Section: Experiments a Experimental Set-up A: Datasetsmentioning
confidence: 99%
“…For the evaluation metrics, we use average accuracy (A avg ) in all task transitions (excluding the first task's accuracy as it is not incrementally learned [9], [20]), final model's accuracy (A k , where k is the final or last task. It is denoted as 'Average' in [44]), forgetting (F k ), intransigence (I k ) and multi-head intransigence (I mh k ) metrics at task t k proposed by [3].…”
Section: D: Evaluation Metricsmentioning
confidence: 99%
See 2 more Smart Citations