2020
DOI: 10.48550/arxiv.2006.07862
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Exploiting Higher Order Smoothness in Derivative-free Optimization and Continuous Bandits

Abstract: We study the problem of zero-order optimization of a strongly convex function. The goal is to find the minimizer of the function by a sequential exploration of its values, under measurement noise. We study the impact of higher order smoothness properties of the function on the optimization error and on the cumulative regret. To solve this problem we consider a randomized approximation of the projected gradient descent algorithm. The gradient is estimated by a randomized procedure involving two function evaluat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
12
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(14 citation statements)
references
References 13 publications
2
12
0
Order By: Relevance
“…All these results are similar to the rates as in [NS17]. In fact, these rates are sharper than the noise-free rates as provided in [APT20], where higher-order smoothness -although for the perturbed setting -is explicitly taken into account. The key difference with all of the aforementioned works is that we can select our smoothing parameter as being for example of the form δ k = δ/(1 + k) without the fear of numerical problems.…”
Section: Numerical Considerationssupporting
confidence: 83%
“…All these results are similar to the rates as in [NS17]. In fact, these rates are sharper than the noise-free rates as provided in [APT20], where higher-order smoothness -although for the perturbed setting -is explicitly taken into account. The key difference with all of the aforementioned works is that we can select our smoothing parameter as being for example of the form δ k = δ/(1 + k) without the fear of numerical problems.…”
Section: Numerical Considerationssupporting
confidence: 83%
“…2. Note that the results of this work have better dependency ε(N ) or N (ε) than Gasnikov's one-point method only if β > 2 else another technique in Theorem 1 is better (see [8] or Theorem 5.1 in [1]). The result in this work is achieved using both kernel smoothing technique and measure concentration inequalities.…”
Section: Contributionsmentioning
confidence: 82%
“…3. The lower bound for strongly convex case is got under conditions γ ≥ N − 1 /2+ 1 /β (otherwise it is better to use convex methods) and (see [1]) 2γ ≤ max x∈Q ∇f (x) .…”
Section: Contributionsmentioning
confidence: 99%
See 1 more Smart Citation
“…Gradient-free methods. Let us highlight the main works devoted to the zeroth-order methods: for two-point feedback [19,17,7,11], for one-point feedback [2,9,1,18].…”
Section: Related Workmentioning
confidence: 99%