2004
DOI: 10.1007/978-3-540-27819-1_46
|View full text |Cite
|
Sign up to set email alerts
|

The Budgeted Multi-armed Bandit Problem

Abstract: The following coins problem is a version of a multi-armed bandit problem where one has to select from among a set of objects, say classifiers, after an experimentation phase that is constrained by a time or cost budget. The question is how to spend the budget. The problem involves pure exploration only, differentiating it from typical multi-armed bandit problems involving an exploration/exploitation tradeoff [BF85]. It is an abstraction of the following scenarios: choosing from among a set of alternative treat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
27
0
1

Year Published

2006
2006
2021
2021

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 31 publications
(28 citation statements)
references
References 1 publication
0
27
0
1
Order By: Relevance
“…Figure 3 supports the predictiveness of (12) with respect to relative empirical costs. Note that, if one assumes the same target confidence τ for all bandits b i ∈ B, then accounting for τ inĉ i would not affect the ordering of bandits according tô c i , as an "easier" bandit according to (12) would also have a smaller expected completion cost for any value of τ . By using (12), we also ignore the effort previously expended on a given bandit.…”
Section: Bayesian Greedy Confidence Pursuitmentioning
confidence: 99%
See 4 more Smart Citations
“…Figure 3 supports the predictiveness of (12) with respect to relative empirical costs. Note that, if one assumes the same target confidence τ for all bandits b i ∈ B, then accounting for τ inĉ i would not affect the ordering of bandits according tô c i , as an "easier" bandit according to (12) would also have a smaller expected completion cost for any value of τ . By using (12), we also ignore the effort previously expended on a given bandit.…”
Section: Bayesian Greedy Confidence Pursuitmentioning
confidence: 99%
“…While considering the number of trials already spent on a bandit could improve on the performance of (12), it would require steps to avoid the "sunk-cost" fallacy of economics, as manifested by premature commitment to bandits wrongly identified as "easy". (12). From darkest to lightest the points represent selecting the top 1, 3, and 5 arms of a 10-armed bandit.…”
Section: Bayesian Greedy Confidence Pursuitmentioning
confidence: 99%
See 3 more Smart Citations