2019
DOI: 10.48550/arxiv.1908.02983
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pseudo-Labeling and Confirmation Bias in Deep Semi-Supervised Learning

Abstract: Semi-supervised learning, i.e. jointly learning from labeled an unlabeled samples, is an active research topic due to its key role on relaxing human annotation constraints. In the context of image classification, recent advances to learn from unlabeled samples are mainly focused on consistency regularization methods that encourage invariant predictions for different perturbations of unlabeled samples. We, conversely, propose to learn from unlabeled data by generating soft pseudolabels using the network predict… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
82
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 34 publications
(84 citation statements)
references
References 19 publications
0
82
1
Order By: Relevance
“…Additional, dependent on the accuracy of the pseudo-labels, we increase the amount of labelled data the model has access to and reduce overfitting to the initally small label set. There are many ways to incorporate unlabelled data / pseudo-label pairs into the loss function but the most common ways are to either create a specific loss term for the unlabelled data pseudo-label pairs [12], [18] or by using composite batches containing both labelled and unlabelled data and keeping the standard supervised classification loss [20], [33].…”
Section: B Pseudo-labelling Techniquesmentioning
confidence: 99%
See 4 more Smart Citations
“…Additional, dependent on the accuracy of the pseudo-labels, we increase the amount of labelled data the model has access to and reduce overfitting to the initally small label set. There are many ways to incorporate unlabelled data / pseudo-label pairs into the loss function but the most common ways are to either create a specific loss term for the unlabelled data pseudo-label pairs [12], [18] or by using composite batches containing both labelled and unlabelled data and keeping the standard supervised classification loss [20], [33].…”
Section: B Pseudo-labelling Techniquesmentioning
confidence: 99%
“…As pointed out by Arazo et al [20] there is a potential pitfall in this style of approach. Networks are often wrong and the neural network can overfit to its own incorrectly guessed pseudo-labels in a process termed confirmation bias.…”
Section: B Pseudo-labelling Techniquesmentioning
confidence: 99%
See 3 more Smart Citations