2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412716
|View full text |Cite
|
Sign up to set email alerts
|

Rethinking deep active learning: Using unlabeled data at model training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
24
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 36 publications
(25 citation statements)
references
References 12 publications
0
24
0
Order By: Relevance
“…We call this the high budget regime, while the regime characterized by failure (or "cold start") is called the low budget regime. In the low budget regime, it has been shown that random selection outperforms most querying strategies [see Attenberg and Provost, 2010, Mittal et al, 2019, Zhu et al, 2020, Siméoni et al, 2021, Chandra et al, 2021. Different accounts have been offered to explain this observation: (i) failure to model uncertainty, which is more severe with a small budget of labels [Nguyen et al, 2015, Gal andGhahramani, 2016]; (ii) difficulty to accomplish diversity in high dimensional input spaces, where traditional metrics (such as the Euclidean distance) are not meaningful.…”
Section: Introductionmentioning
confidence: 99%
“…We call this the high budget regime, while the regime characterized by failure (or "cold start") is called the low budget regime. In the low budget regime, it has been shown that random selection outperforms most querying strategies [see Attenberg and Provost, 2010, Mittal et al, 2019, Zhu et al, 2020, Siméoni et al, 2021, Chandra et al, 2021. Different accounts have been offered to explain this observation: (i) failure to model uncertainty, which is more severe with a small budget of labels [Nguyen et al, 2015, Gal andGhahramani, 2016]; (ii) difficulty to accomplish diversity in high dimensional input spaces, where traditional metrics (such as the Euclidean distance) are not meaningful.…”
Section: Introductionmentioning
confidence: 99%
“…To address the first problem, researchers have considered using generative networks for data augmentation [162] or assigning pseudo-labels to high-confidence samples in order to expand the labeled training set [166]. Some researchers have also used labeled and unlabeled datasets to combine supervised and semisupervised training across AL cycles [65,148]. In addition, previous heuristicbased AL [139] query strategies have proven to be ineffective when applied to DL [138]; therefore, for the one-by-one query strategy in classic AL, many researchers focus on the improvement of the batch sample query strategy [10,51,84,183], taking both the amount of information and the diversity of batch samples into account.…”
Section: The Necessity and Challenge Of Combining DL And Almentioning
confidence: 99%
“…Previous works have successfully managed to integrate unlabeled data into AL using self-supervised learning and semi-supervised learning. Siméoni et al [20] showed that initializing the target model with the features obtained from self-supervised pretraining gives AL a kickstart in performance. Contemporaneously, Mottaghi & Yeung [7] also used this technique in combination with a GAN based AL method and reported SOTA results on SVHN, CIFAR10, ImageNet, CelebA datasets.…”
Section: Related Workmentioning
confidence: 99%