2017
DOI: 10.1073/pnas.1611835114
|View full text |Cite
|
Sign up to set email alerts
|

Overcoming catastrophic forgetting in neural networks

Abstract: The ability to learn tasks in a sequential fashion is crucial to the development of artificial intelligence. Until now neural networks have not been capable of this and it has been widely thought that catastrophic forgetting is an inevitable feature of connectionist models. We show that it is possible to overcome this limitation and train networks that can maintain expertise on tasks that they have not experienced for a long time. Our approach remembers old tasks by selectively slowing down learning on the wei… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

22
3,870
1
8

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 4,564 publications
(4,267 citation statements)
references
References 36 publications
22
3,870
1
8
Order By: Relevance
“…Similarly, various properties can be derived from a single stream, like visual aspects (depth, figure-ground separation, segmentation), from an object recognition stream, where each aspect sub-stream is learned via a separate cost function. For tasks sharing outputs, and thus having overlap over different tasks, evidence increasingly suggests that the brain selectively "protects" synapses for modification by new tasks, effectively "unsharing" these parameters between tasks (Kirkpatrick et al, 2017).…”
Section: Multiple Pathways As a Solution For Catastrophic Forgettingmentioning
confidence: 99%
“…Similarly, various properties can be derived from a single stream, like visual aspects (depth, figure-ground separation, segmentation), from an object recognition stream, where each aspect sub-stream is learned via a separate cost function. For tasks sharing outputs, and thus having overlap over different tasks, evidence increasingly suggests that the brain selectively "protects" synapses for modification by new tasks, effectively "unsharing" these parameters between tasks (Kirkpatrick et al, 2017).…”
Section: Multiple Pathways As a Solution For Catastrophic Forgettingmentioning
confidence: 99%
“…Such a weak causality indicates novelty of information, and thus encourages the brain to employ knowledge about past experiences in its learning of the new information. Artificial neural networks, on the other hand, face difficulties in this task, because new information will interfere with previously learned knowledge [30]. Though deep learning excels in finding intricate structures in big data sets, it does less well than the brain when only small quantities of training data are available.…”
Section: Brain-inspired Computingmentioning
confidence: 99%
“…However, current neural network approaches have still not been able to fully implement continual learning, and there is also inevitable catastrophic forgetting associated with this mode of learning. In an attempt to enable agents to continuously learn without catastrophic forgetting, James et al [26] recently proposed training of networks, using an elastic weight consolidation (EWC) algorithm, that can maintain expertise from previously learnt tasks. For gradient policy learning, a deep deterministic-policy gradient (DDPG) algorithm [27] has been proposed to continuously improve policy, whilst an agent explores its environment.…”
Section: Introductionmentioning
confidence: 99%
“…As noted earlier, current GPS approaches can only handle scenarios where data from all tasks is simultaneously made available during the early training stage, which constitutes an impractical constraint for consecutive task learning. By exploiting and adapting the recently developed EWC algorithm [26], we propose incorporation of Fisher information, to protect weights that are important for previous tasks, while learning the new task at hand. To some extent, this also overcomes catastrophic forgetting, in our proposed approach to sequential multi-task learning.…”
Section: Introductionmentioning
confidence: 99%