2013
DOI: 10.1007/978-3-642-40988-2_16
|View full text |Cite
|
Sign up to set email alerts
|

Greedy Confidence Pursuit: A Pragmatic Approach to Multi-bandit Optimization

Abstract: Abstract. We address the practical problem of maximizing the number of high-confidence results produced among multiple experiments sharing an exhaustible pool of resources. We formalize this problem in the framework of bandit optimization as follows: given a set of multiple multi-armed bandits and a budget on the total number of trials allocated among them, select the top-m arms (with high confidence) for as many of the bandits as possible. To solve this problem, which we call greedy confidence pursuit, we dev… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 8 publications
(18 reference statements)
0
6
0
Order By: Relevance
“…Consistency-based regularization (Berthelot et al 2019b;Xie et al 2020;Laine and Aila 2017;Berthelot et al 2019a;Tarvainen and Valpola 2017) applies a consistency loss by enforcing invariance on unlabeled data under different augmentations. Pseudo-labeling relies on the model's high confident predictions to produce pseudo-labels (Lee et al 2013;Bachman, Alsharif, and Precup 2014;Arazo et al 2020) for unlabeled data and trains them jointly with labeled data. FixMatch (Sohn et al 2020a) is a combination of both consistency-based regularization and pseudolabeling approaches.…”
Section: Related Workmentioning
confidence: 99%
“…Consistency-based regularization (Berthelot et al 2019b;Xie et al 2020;Laine and Aila 2017;Berthelot et al 2019a;Tarvainen and Valpola 2017) applies a consistency loss by enforcing invariance on unlabeled data under different augmentations. Pseudo-labeling relies on the model's high confident predictions to produce pseudo-labels (Lee et al 2013;Bachman, Alsharif, and Precup 2014;Arazo et al 2020) for unlabeled data and trains them jointly with labeled data. FixMatch (Sohn et al 2020a) is a combination of both consistency-based regularization and pseudolabeling approaches.…”
Section: Related Workmentioning
confidence: 99%
“…To alleviate this problem, noise can be added to the model during the inference time to generate more accurate predictions. This method is used in the Pseudo-Ensemble Agreement [48] and has demonstrated excellent performance. Thus, a teacher model injected with noise can be inferred to generate more precise targets than that not injected with noise.…”
Section: Data Augmentationmentioning
confidence: 99%
“…Consistency based SSL has been extensively studied in the context of deep learning in recent years [10,[14][15][16]25] . These methods leverage unlabeled data by adding an unsupervised regularization term to the standard supervised loss:…”
Section: Related Workmentioning
confidence: 99%