2011
DOI: 10.1007/978-3-642-24412-4_14
|View full text |Cite
|
Sign up to set email alerts
|

Lipschitz Bandits without the Lipschitz Constant

Abstract: Abstract. We consider the setting of stochastic bandit problems with a continuum of arms indexed by [0, 1] d . We first point out that the strategies considered so far in the literature only provided theoretical guarantees of the form: given some tuning parameters, the regret is small with respect to a class of environments that depends on these parameters. This is however not the right perspective, as it is the strategy that should adapt to the specific bandit environment at hand, and not the other way roun… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
58
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 51 publications
(62 citation statements)
references
References 10 publications
(4 reference statements)
4
58
0
Order By: Relevance
“…In terms of n, it nearly matches 2 the (n 1+k 2+k ) lower bound [13], for k-variate Lipschitz continuous mean reward functions. As explained earlier, the per-round regret R(n)/n approaches zero (as n increases), at a rate exponential in k. Thus for k d, we avoid the curse of dimensionality.…”
Section: K D) = O(poly(k) · O(log D))supporting
confidence: 59%
See 3 more Smart Citations
“…In terms of n, it nearly matches 2 the (n 1+k 2+k ) lower bound [13], for k-variate Lipschitz continuous mean reward functions. As explained earlier, the per-round regret R(n)/n approaches zero (as n increases), at a rate exponential in k. Thus for k d, we avoid the curse of dimensionality.…”
Section: K D) = O(poly(k) · O(log D))supporting
confidence: 59%
“…Say the linear parameter matrix A, or even the sub-space spanned by its rows, was known. We then know a lower bound of (n 1+k 2+k ) on regret, for k-variate Lipschitz continuous mean rewards [13]. In terms of n, our bound nearly matches this lower bound, albeit for a slightly restricted class of Lipschitz continuous mean reward functions.…”
Section: Theorem 2 For Anymentioning
confidence: 53%
See 2 more Smart Citations
“…As noted by Hansen and Jaumard (1995), it is unclear if this approach provides any advantage, considering that other successful heuristics are already available. This argument still applies to this day to relatively recent algorithms such as those by Kvasov et al (2003) and Bubeck, Stoltz, and Yu (2011).…”
Section: Global Optimization On Black-box Functionmentioning
confidence: 98%