2022
DOI: 10.48550/arxiv.2210.02482
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fisher information lower bounds for sampling

Abstract: We prove two lower bounds for the complexity of non-log-concave sampling within the framework of , who introduced the use of Fisher information (FI) bounds as a notion of approximate first-order stationarity in sampling. Our first lower bound shows that averaged LMC is optimal for the regime of large FI by reducing the problem of finding stationary points in non-convex optimization to sampling. Our second lower bound shows that in the regime of small FI, obtaining a FI of at most ε 2 from the target distributi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 16 publications
(24 reference statements)
0
4
0
Order By: Relevance
“…5 At the time of writing, there was no setting in which current mixing bounds for the (Unadjusted) Langevin Algorithm were known to be optimal. The concurrent work [23] also shows an optimality result. Their result concerns an entirely different setting (non-convex potentials where error is measured in Fisher information) and applies to a variant of the Langevin Algorithm (that outputs the averaged iterate) in the specific regime of very large mixing error (namely ε ≈ √ d, where d is the dimension).…”
Section: Related Workmentioning
confidence: 74%
See 2 more Smart Citations
“…5 At the time of writing, there was no setting in which current mixing bounds for the (Unadjusted) Langevin Algorithm were known to be optimal. The concurrent work [23] also shows an optimality result. Their result concerns an entirely different setting (non-convex potentials where error is measured in Fisher information) and applies to a variant of the Langevin Algorithm (that outputs the averaged iterate) in the specific regime of very large mixing error (namely ε ≈ √ d, where d is the dimension).…”
Section: Related Workmentioning
confidence: 74%
“…An important but difficult question for bridging theory and practice is to understand to what extent fast mixing carries over to families of target distributions π under weaker assumptions than log-concavity (i.e., weaker assumptions on potentials than convexity). There have been several recent breakthrough results towards this direction under various assumptions like dissipativity or isoperimetry or under weaker notions of convergence (see the previous work section §1.3 and the very recent papers [6,23]). However, on one hand tight mixing bounds are unknown for any of these settings (see Footnote 5), and on the other hand the non-convexity assumptions made in the sampling literature are often markedly different from those in the optimization literature.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Balasubramanian et al (2022) show that even for non-log-concave distributions, averaged Langevin Monte Carlo converges quickly to a distribution with low relative Fisher information to the target distribution, although this does not imply that the distribution is close to the target distribution with respect to other measures such as the total variation distance. Chewi et al (2022a) prove corresponding lower bounds. Woodard et al (2009) show that the mixing time of parallel and simulated tempering for certain distributions can scale exponentially with d, but in a setting different from ours.…”
Section: Related Workmentioning
confidence: 93%