Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence 2017
DOI: 10.24963/ijcai.2017/350
|View full text |Cite
|
Sign up to set email alerts
|

Thresholding Bandits with Augmented UCB

Abstract: In this paper we propose the Augmented-UCB (AugUCB) algorithm for a fixed-budget version of the thresholding bandit problem (TBP), where the objective is to identify a set of arms whose quality is above a threshold. A key feature of AugUCB is that it uses both mean and variance estimates to eliminate arms that have been sufficiently explored; to the best of our knowledge this is the first algorithm to employ such an approach for the considered TBP. Theoretically, we obtain an upper bound on the loss (probabili… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 13 publications
(18 reference statements)
0
10
0
Order By: Relevance
“…(Chen et al, 2014) also develops the CSAR algorithm for the fixed-budget setting which can also be used for TBP. The result was improved by recent followup work (Locatelli et al, 2016;Mukherjee et al, 2017) under the fixed budget setting. Chen et al (2015) considered TBP in the context of budget allocation for crowdsourced classification in the Bayesian framework.…”
Section: Introductionmentioning
confidence: 87%
See 1 more Smart Citation
“…(Chen et al, 2014) also develops the CSAR algorithm for the fixed-budget setting which can also be used for TBP. The result was improved by recent followup work (Locatelli et al, 2016;Mukherjee et al, 2017) under the fixed budget setting. Chen et al (2015) considered TBP in the context of budget allocation for crowdsourced classification in the Bayesian framework.…”
Section: Introductionmentioning
confidence: 87%
“…We note that previous works on TBP in the fixed budget setting (Locatelli et al, 2016;Mukherjee et al, 2017) cannot be implemented in our fixed-confidence setting.…”
Section: Baselines and Implementation Detailsmentioning
confidence: 96%
“…We consider three settings named Threshold 1-3, which are based on Experiment 1-2 in Locatelli et al (2016) and Experiment 4 in Mukherjee et al (2017).…”
Section: Threshold Settingsmentioning
confidence: 99%
“…This variant of the multi-armed bandit problem was introduced by Locatelli et al (2016), who provided an algorithm for solving the problem with matching upper and lower bounds. Mukherjee et al (2017) and Zhong et al (2017) have since provided algorithmic extensions that incorporate variance estimates and provide guarantees in asynchronous settings.…”
Section: Thresholding Banditsmentioning
confidence: 99%