2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.01322
|View full text |Cite
|
Sign up to set email alerts
|

Maintaining Discrimination and Fairness in Class Incremental Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
186
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 294 publications
(206 citation statements)
references
References 13 publications
1
186
0
1
Order By: Relevance
“…Xiang et al [ 78 ] proposed an algorithm based on dynamic correction vectors to solve the deviation from knowledge distillation and model overfitting problems. Zhao et al [ 79 ] combined weight adjustment and knowledge distillation in order to balance the new and old knowledge. Javed et al [ 80 ] proposed a dynamic threshold shift method to improve the limitations of the deviation in a general knowledge distillation model.…”
Section: Methods Descriptionmentioning
confidence: 99%
“…Xiang et al [ 78 ] proposed an algorithm based on dynamic correction vectors to solve the deviation from knowledge distillation and model overfitting problems. Zhao et al [ 79 ] combined weight adjustment and knowledge distillation in order to balance the new and old knowledge. Javed et al [ 80 ] proposed a dynamic threshold shift method to improve the limitations of the deviation in a general knowledge distillation model.…”
Section: Methods Descriptionmentioning
confidence: 99%
“…Rehearsal / replay consists in replaying old data to the model at each new training step. One way is to select some samples from the incoming data stream and store them inside an episodic memory [4], [19], [22], [26]. Rehearsal is currently the most successful strategy to counter forgetting.…”
Section: A Class-incremental Learningmentioning
confidence: 99%
“…Regularization is a strategy consisting in implementing a protectionist policy on learned knowledge. Usually, the regularization is directly applied on the output layer using knowledge distillation [16], [19], [26]: the previous state of the model is used as a teacher for the new model in order to maintain the discrimination between old classes when optimizing new outputs. Other methods add the regularization on every weights of the model [1], [12], [24].…”
Section: A Class-incremental Learningmentioning
confidence: 99%
“…Regularization methods add specific regularization terms to consolidate knowledge learned before. Li and Hoiem (2017) introduced the knowledge distillation (Hinton et al, 2015) to penalize model logit change, and it has been widely employed in Rebuffi et al (2017);Castro et al (2018); ; Hou et al (2019); Zhao et al (2019). Another direction is to regularize parameters crucial to old knowledge according to various importance measures (Kirkpatrick et al, 2017;Zenke et al, 2017;Aljundi et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Exemplar replay methods store past samples, a.k.a exemplars, and replay them periodically. Instead of selecting exemplars at random, Rebuffi et al (2017) incorporated the Herding technique (Welling, 2009) to choose exemplars that best approximate the mean feature vector of a class, and it is widely used in Castro et al (2018); ; Hou et al (2019); Zhao et al (2019); Mi et al (2020a,b). Ramalho and Garnelo (2019) proposed to store samples that the model is least confident.…”
Section: Related Workmentioning
confidence: 99%