2021 IEEE International Symposium on Information Theory (ISIT) 2021
DOI: 10.1109/isit45174.2021.9518176
|View full text |Cite
|
Sign up to set email alerts
|

Regret Bounds for Safe Gaussian Process Bandit Optimization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 5 publications
0
6
0
Order By: Relevance
“…The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) variance, the BQ noisy lower bound is known to be ⌦(T 1 2 ) (Plaskota 1996;Cai, Lam, and Scarlett 2023), and the BO noisy lower bound is known as ✏ = ⌦(T ⌫ 2⌫+d ) (Scarlett, Bogunovic, and Cevher 2017;Cai and Scarlett 2021). To compare against these, for the first dot point in Theorem 2, we consider the following specific c values:…”
Section: Lower Boundsmentioning
confidence: 99%
“…The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) variance, the BQ noisy lower bound is known to be ⌦(T 1 2 ) (Plaskota 1996;Cai, Lam, and Scarlett 2023), and the BO noisy lower bound is known as ✏ = ⌦(T ⌫ 2⌫+d ) (Scarlett, Bogunovic, and Cevher 2017;Cai and Scarlett 2021). To compare against these, for the first dot point in Theorem 2, we consider the following specific c values:…”
Section: Lower Boundsmentioning
confidence: 99%
“…This is a quite mild assumption since it only requires that one can find a probability distribution over the set of actions under which the expected cost is less than a strictly negative value. This is in sharp constraint to existing KB algorithms for hard constraints that typically require the existence of an initial safe action (Sui et al, 2018;Amani et al, 2020).…”
Section: Problem Formulation and Preliminariesmentioning
confidence: 99%
“…To this end, there have been exciting recent advances in the theoretical analysis of constrained kernelized bandits. In particular, (Sui et al, 2015;Berkenkamp et al, 2016;Sui et al, 2018) propose algorithms with convergence guarantees, while (Amani et al, 2020), to the best our knowledge, is the first work that establishes regret bounds for their developed algorithm, although under the Bayesian-type 1 setting. These algorithms mainly focus on KB with a hard constraint such as safety, i.e., the selected action in each round needs to satisfy the constraint with a high probability.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The only known results on safe exploration in multi-armed bandits address the case with continuous, convex arm spaces and convex constraints. The learner can converge to the optimal solution in these settings without violating the constraints [16,17]. Conversely, the case with discrete and/or non-convex arm spaces or non-convex constraints, such as ours, is unexplored in the literature so far.…”
Section: Related Workmentioning
confidence: 99%