2022
DOI: 10.48550/arxiv.2201.08518
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Optimal variance-reduced stochastic approximation in Banach spaces

Abstract: We study the problem of estimating the fixed point of a contractive operator defined on a separable Banach space. Focusing on a stochastic query model that provides noisy evaluations of the operator, we analyze a variance-reduced stochastic approximation scheme, and establish non-asymptotic bounds for both the operator defect and the estimation error, measured in an arbitrary semi-norm. In contrast to worst-case guarantees, our bounds are instance-dependent, and achieve the local asymptotic minimax risk non-as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(9 citation statements)
references
References 27 publications
0
9
0
Order By: Relevance
“…The growth condition requires the incremental update H(x, ξ) to grow at most linearly in both x and a non-negative function g : Ξ → R that captures the contribution of data ξ to the norm growth of H(x, ξ) . It would be emphasized that we assume {g(ξ t )} t≥0 has uniformly bounded p-th moments, much milder than previous almost surely uniformly boundedness [Chen et al, 2021c, Doan et al, 2020, Mou et al, 2022a.…”
Section: Consistency Guaranteementioning
confidence: 99%
See 2 more Smart Citations
“…The growth condition requires the incremental update H(x, ξ) to grow at most linearly in both x and a non-negative function g : Ξ → R that captures the contribution of data ξ to the norm growth of H(x, ξ) . It would be emphasized that we assume {g(ξ t )} t≥0 has uniformly bounded p-th moments, much milder than previous almost surely uniformly boundedness [Chen et al, 2021c, Doan et al, 2020, Mou et al, 2022a.…”
Section: Consistency Guaranteementioning
confidence: 99%
“…A sufficient condition for (3) is almost surely Lipschitz continuity, meaning that |H(x, ξ) − H(y, ξ)| ≤ L H |x − y| holds for any x, y ∈ R d and ξ ∈ Ξ. This type of condition is commonly used in machine learning, as demonstrated by the A2 condition in [Mou et al, 2022a].…”
Section: Assumption 1 (Local Linearity) There Exist Constantsmentioning
confidence: 99%
See 1 more Smart Citation
“…Consequently, existing onpolicy first-order policy optimization methods require restrictive assumptions, e.g., all the iterates of policies are random enough, which is problematic when the optimal policy does not possess this structure (e.g., being deterministic). In spite of intensive research effort of on-policy evaluation (Tsitsiklis and Van Roy, 1999;Yu and Bertsekas, 2009;Zhang et al, 2021b;Mou et al, 2022), one seemly unresolved problem in RL is whether one can design sampling-efficient on-policy evaluation algorithms for insufficiently random policies and use them for policy optimization (see Remark 1 of Lan, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…In addition, they establish convergence analysis in 2 -norm, which is not a natural metric for the underlying problem, thus leading to worse dependence on other problem parameters, i.e., dimension of the transition kernel. It is also noteworthy that recent work by Mou et al (2022) proposed a variance-reduced stochastic approximation approach that solves the AMDP policy evaluation problem in span semi-norm under the generative model. However, their results do not directly extend to the Markovian noise setting.…”
mentioning
confidence: 99%