2020
DOI: 10.1007/978-3-030-58536-5_31
|View full text |Cite
|
Sign up to set email alerts
|

GDumb: A Simple Approach that Questions Our Progress in Continual Learning

Abstract: Standardized benchmarks drive progress in machine learning. However, with repeated testing, the risk of overfitting grows as algorithms overexploit benchmark idiosyncrasies. In our work, we seek to mitigate this challenge by compiling ever-expanding large-scale benchmarks called Lifelong Benchmarks. As exemplars of our approach, we create Lifelong-CIFAR10 and Lifelong-ImageNet, containing (for now) 1.69M and 1.98M test samples, respectively. While reducing overfitting, lifelong benchmarks introduce a key chall… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
283
1

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 313 publications
(317 citation statements)
references
References 71 publications
(67 reference statements)
0
283
1
Order By: Relevance
“…The latter is considered a strong CL baseline (Maltoni and Lomonaco, 2019) and thus we use this approach in this study. We also compare against a variation of the replay-based GDumb baseline (Prabhu et al, 2020). GDumb collects examples into a memory buffer with a limited budget size k, balancing the distribution over labels by greedily sampling underrepresented label types and ejecting over-sampled label types.…”
Section: Continuous Learningmentioning
confidence: 99%
“…The latter is considered a strong CL baseline (Maltoni and Lomonaco, 2019) and thus we use this approach in this study. We also compare against a variation of the replay-based GDumb baseline (Prabhu et al, 2020). GDumb collects examples into a memory buffer with a limited budget size k, balancing the distribution over labels by greedily sampling underrepresented label types and ejecting over-sampled label types.…”
Section: Continuous Learningmentioning
confidence: 99%
“…The approach in [28] tries to retrieve only the samples that are most conflicted. At the same time, [29] proposes to improve performance by greedily storing samples in memory and retraining on these stored samples while testing. The work in [30] proposes an expansion-based approach for task-free continual learning built upon the Bayesian nonparametric.…”
Section: A Incremental Learningmentioning
confidence: 99%
“…Comparison is also performed with [27], which proposes an intermediate expert to adapt the target model to the new task and [28], which retrieves the samples that are frequently conflicted. We also include our comparison with [29], which greedily stores samples in memory and trains a model from scratch and uses these samples during testing. Finally, we compare with [30], which increases the number of neural network experts under the Bayesian non-parametric framework We carried out experiments for comparing the results obtained from the proposed approach with different algorithms: 1.…”
Section: Comparison With Existing Literaturementioning
confidence: 99%
See 1 more Smart Citation
“…Our work follows this transfer learning paradigm but our main focus is to investigate the regression phenomenon when updating backbone pre-trained models. Another related stream of research is lifelong learning (Lopez-Paz and Ranzato, 2017;Yoon et al, 2018;Delange et al, 2021;Sun et al, 2019;Chuang et al, 2020), incremental learning (Rebuffi et al, 2017Chaudhry et al, 2018;Prabhu et al, 2020), or concept drifting (Schlimmer and Granger, 1986;Tsymbal, 2004;Klinkenberg, 2005;Žliobaitė I., 2016) which aims to accumulate knowledge learned either in previous tasks or from data with changing distribution. The model update regression problem differs in that models are trained on the same task and dataset, but we update from one model to another.…”
Section: Transfer Learning Lifelong Learning and Concept Driftingmentioning
confidence: 99%