2021
DOI: 10.48550/arxiv.2111.07928
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Target Layer Regularization for Continual Learning Using Cramer-Wold Generator

Abstract: We propose an effective regularization strategy (CW-TaLaR) for solving continual learning problems. It uses a penalizing term expressed by the Cramer-Wold distance between two probability distributions defined on a target layer of an underlying neural network that is shared by all tasks, and the simple architecture of the Cramer-Wold generator for modeling output data representation. Our strategy preserves target layer distribution while learning a new task but does not require remembering previous tasks' data… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 8 publications
0
2
0
Order By: Relevance
“…2). We follow the domain incremental setup (van de Ven et al, 2020;Kessler et al, 2021;Raghavan & Balaprakash, 2021;Hsu et al, 2018;Mazur et al, 2021;He & Zhu, 2022), where the number of classes is constant and task information is not available, but we also compare against task-incremental methods (Zenke et al, 2017;Shin et al, 2017;Nguyen et al, 2017;Kirkpatrick et al, 2017;Li & Hoiem, 2017;Rao et al, 2019;Sokar et al, 2021;Yoon et al, 2018;Han & Guo, 2021;Chaudhry et al, 2021 in Appendix C Tables 3 and 4, where the task information is available, to give a complete overview of current methods. The different Continual Learning setups are given in Fig.…”
Section: Continual Supervised Learning Scenariosmentioning
confidence: 99%
See 1 more Smart Citation
“…2). We follow the domain incremental setup (van de Ven et al, 2020;Kessler et al, 2021;Raghavan & Balaprakash, 2021;Hsu et al, 2018;Mazur et al, 2021;He & Zhu, 2022), where the number of classes is constant and task information is not available, but we also compare against task-incremental methods (Zenke et al, 2017;Shin et al, 2017;Nguyen et al, 2017;Kirkpatrick et al, 2017;Li & Hoiem, 2017;Rao et al, 2019;Sokar et al, 2021;Yoon et al, 2018;Han & Guo, 2021;Chaudhry et al, 2021 in Appendix C Tables 3 and 4, where the task information is available, to give a complete overview of current methods. The different Continual Learning setups are given in Fig.…”
Section: Continual Supervised Learning Scenariosmentioning
confidence: 99%
“…We achieve comparable results to current state-of-the-art approaches (see Tables 1 and 2) on all three supervised learning benchmarks. (Raghavan & Balaprakash, 2021) 98.71 (± 0.06) 97.51 (± 0.05) Target layer regularization (Mazur et al, 2021) 80.64 (± 1.25) Natural continual learning (Kao et al, 2021) 38.79 (± 0.24) Target layer regularization (Mazur et al, 2021) 74.89 (± 0.61 Memory aware synapses (He & Zhu, 2022) 73.50 (± 1.54)…”
Section: Continual Supervised Learning Scenariosmentioning
confidence: 99%