2022
DOI: 10.48550/arxiv.2207.12061
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Balancing Stability and Plasticity through Advanced Null Space in Continual Learning

Abstract: Continual learning is a learning paradigm that learns tasks sequentially with resources constraints, in which the key challenge is stability-plasticity dilemma, i.e., it is uneasy to simultaneously have the stability to prevent catastrophic forgetting of old tasks and the plasticity to learn new tasks well. In this paper, we propose a new continual learning approach, Advanced Null Space (AdNS), to balance the stability and plasticity without storing any old data of previous tasks. Specifically, to obtain bette… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 30 publications
(69 reference statements)
0
2
0
Order By: Relevance
“…NCL (Kao et al, 2021) combines the idea of gradient projection and Bayesian weight regularization to mitigate catastrophic forgetting. In spite of minimizing backward interference, these approaches suffer poor forward knowledge transfer and lack plasticity (Kong et al, 2022). TRGP (Lin et al, 2022) expands the model with trust regions to achieve better performance on new tasks by introducing additional scale parameters.…”
Section: Gradient Projection Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…NCL (Kao et al, 2021) combines the idea of gradient projection and Bayesian weight regularization to mitigate catastrophic forgetting. In spite of minimizing backward interference, these approaches suffer poor forward knowledge transfer and lack plasticity (Kong et al, 2022). TRGP (Lin et al, 2022) expands the model with trust regions to achieve better performance on new tasks by introducing additional scale parameters.…”
Section: Gradient Projection Methodsmentioning
confidence: 99%
“…To mitigate forgetting and facilitate forward knowledge transfer, replay-based methods (Lopez-Paz & Ranzato, 2017;Shin et al, 2017;Choi et al, 2021) stores some old samples in the memory, and expansion-based methods (Rusu et al, 2016;Yoon et al, 2017;2019) expand the model structure to accommodate incoming knowledge. However, these methods require either extra memory buffers (Parisi et al, 2019) or a growing network architecture as new tasks continually arrive (Kong et al, 2022), which are always computationally expansive (De Lange et al, 2021). Thus, promoting performance within a fixed network capacity remains challenging.…”
Section: Work In Progressmentioning
confidence: 99%