2023
DOI: 10.1109/tpami.2022.3163806
|View full text |Cite
|
Sign up to set email alerts
|

Uncertainty-Aware Contrastive Distillation for Incremental Semantic Segmentation

Abstract: A fundamental and challenging problem in deep learning is catastrophic forgetting, i.e. the tendency of neural networks to fail to preserve the knowledge acquired from old tasks when learning new tasks. This problem has been widely investigated in the research community and several Incremental Learning (IL) approaches have been proposed in the past years. While earlier works in computer vision have mostly focused on image classification and object detection, more recently some IL approaches for semantic segmen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 31 publications
(19 citation statements)
references
References 53 publications
0
18
0
Order By: Relevance
“…We apply the constantly updated step-aware weight ψ t ij to the cross-entropy loss based on one-hot pseudo labels ỹt ij in Eq. (12). And the different forgetting paces can be alleviated by re-weighting gradient back-propagation of old classes (see Fig.…”
Section: A Step-aware Gradient Compensationmentioning
confidence: 99%
See 1 more Smart Citation
“…We apply the constantly updated step-aware weight ψ t ij to the cross-entropy loss based on one-hot pseudo labels ỹt ij in Eq. (12). And the different forgetting paces can be alleviated by re-weighting gradient back-propagation of old classes (see Fig.…”
Section: A Step-aware Gradient Compensationmentioning
confidence: 99%
“…It is natural to seek effective incremental semantic segmentation (ISS) methods [7], [8], which can continually learn the model with the training samples of novel classes only. There have been some ISS works [9], [10], [11], [7], [8], [12], [13], which mainly focus on two key challenges, catastrophic forgetting [14] and background shift [9] (see Fig. 1).…”
Section: Introductionmentioning
confidence: 99%
“…3, the first category, known as data-replay methods, involves storing a portion of past training data as exemplar memory such as [26], [27], [28], [29], [30], [31], [32], [33], [34], [35], [36]. The second category, termed datafree methods, includes methods like [37], [38], [39], [40], [41], [42], [43], [44], [45], [46], [47], [48], [49]. These methods utilize transfer learning techniques, such as knowledge distillation, to inherit the capabilities of the old model.…”
Section: Semantic Driftmentioning
confidence: 99%
“…Here we would like to discuss the advantage and necessity of continual learning based on specified models during the period of emerging large models. Although recent [26], [34], [35], [50], [51] Generative-replay Generative-data Generative-feature without storing real data, customized data replay heavy reliance on generative quality, high space complexity [28], [29], [32], [36], [52] Self-supervised Contrastive-learning Pseudo-labeling Foundation-model Driven strong adaptability, exemplar-memory free high training cost, hard to convergence [27], [41], [48], [53], [54] Regularization-based [39], [40], [43], [44], [55] Dynamic-architecture Parameter Allocation Architecture Decomposition Modular Network high model flexibility, high adaptability to diverse data network parameters gradually increases, high space complexity [30], [46], [56], [57], [58] large-model forms [59], [60] achieve fair zero-shot learning ability, they often lack the ability to classify targets with semantic understanding like humans. Another significant concern is cost.…”
Section: Semantic Driftmentioning
confidence: 99%
See 1 more Smart Citation