2021
DOI: 10.48550/arxiv.2103.05478
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Small Errors in Random Zeroth Order Optimization are Imaginary

Abstract: The vast majority of zeroth order optimization methods try to imitate first order methods via some smooth approximation of the gradient. Here, the smaller the smoothing parameter, the smaller the gradient approximation error. We show that for the majority of zeroth order methods this smoothing parameter can however not be chosen arbitrarily small as numerical cancellation errors will dominate. As such, theoretical and numerical performance could differ significantly. Using classical tools from numerical differ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
23
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(23 citation statements)
references
References 23 publications
(35 reference statements)
0
23
0
Order By: Relevance
“…Building upon the work by [KW52; LM67; NY83; ST98; FKM04; NS17], Jongeneel, Yue, and Kuhn [JYK21] show that randomized zeroth-order optimization also benefits from passing to the complex domain as one can derive an inherently numerically stable method, which is in sharp contrast to common finite-difference methods. This work departs from [JYK21] by introducing an indispensable layer of realism; noise.…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…Building upon the work by [KW52; LM67; NY83; ST98; FKM04; NS17], Jongeneel, Yue, and Kuhn [JYK21] show that randomized zeroth-order optimization also benefits from passing to the complex domain as one can derive an inherently numerically stable method, which is in sharp contrast to common finite-difference methods. This work departs from [JYK21] by introducing an indispensable layer of realism; noise.…”
Section: Introductionmentioning
confidence: 99%
“…Optimizers, which based on the context could be globally or locally optimal, are denoted by x . We extend [JYK21] and assume that the objective function f can only be accessed through a zeroth-order oracle that outputs corrupted function evaluations at prescribed test points, that is, with noise. As we only have access to such a zeroth-order oracle, our work belongs to the field of zeroth-order optimization, derivative-free optimization or more generally black-box optimization [CSV09;AH17b].…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations