2020
DOI: 10.1214/19-aos1916
|View full text |Cite
|
Sign up to set email alerts
|

A framework for adaptive MCMC targeting multimodal distributions

Abstract: Please refer to published version for the most recent bibliographic citation information. If a published version is known of, the repository item page linked to above, will contain details on accessing it.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
32
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 35 publications
(32 citation statements)
references
References 54 publications
(93 reference statements)
0
32
0
Order By: Relevance
“…Whereas one typically trains a generative model from a dataset of positive examples, a GFlowNet is trained to match a given energy function and convert it into a sampler, which we view as a generative policy because the composite object s is constructed through a sequence of steps. This is similar to what MCMC methods achieve but does not require a lengthy stochastic search in the space of such objects and avoids the modemixing intractability challenge of MCMC methods (Jasra et al, 2005;Bengio et al, 2013;Pompe et al, 2020). GFlowNets exchange that intractability for the challenge of amortized training of the generative policy.…”
Section: Introductionmentioning
confidence: 66%
See 2 more Smart Citations
“…Whereas one typically trains a generative model from a dataset of positive examples, a GFlowNet is trained to match a given energy function and convert it into a sampler, which we view as a generative policy because the composite object s is constructed through a sequence of steps. This is similar to what MCMC methods achieve but does not require a lengthy stochastic search in the space of such objects and avoids the modemixing intractability challenge of MCMC methods (Jasra et al, 2005;Bengio et al, 2013;Pompe et al, 2020). GFlowNets exchange that intractability for the challenge of amortized training of the generative policy.…”
Section: Introductionmentioning
confidence: 66%
“…Although these methods guarantee that asymptotically (in the length of the chain) we obtain samples drawn from the correct distribution, there is an important set of distributions for which finite chains are unlikely to provide enough diversity of the modes of the distribution. This is known as the mode-mixing problem (Jasra et al, 2005;Bengio et al, 2013;Pompe et al, 2020): the chances of going from one mode to a neighboring one may become exponentially small (and thus require exponentially long chains) if the modes are separated by a long sequence of low-probability configurations. This can be alleviated by burning more computation (sampling longer chains) but becomes exponentially unsustainable with increased mode separation.…”
Section: Contrast With Monte-carlo Markov Chain Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The first collection of approaches uses "mode jumping" proposals, e.g. Tjelmeland and Hegstad [2001], Ibrahim [2009], Liu et al [2000], Tjelmeland and Eidsvik [2004] and Pompe et al [2020]. There is typically no explicit state space augmentation with such methods.…”
Section: Alps Motivationmentioning
confidence: 99%
“…Many methods for accelerating the exploration of MCMC algorithms use prior knowledge of the target. For example, known mode locations may be used to design global moves for the sampler Andricioaei et al (2001); Pompe et al (2020); Lan et al (2014); Sminchisescu and Welling (2007); Sminchisescu et al (2003); Tjelmeland and Hegstad (2001), or moves may be guided by the known derivatives of a differentiable target density Lan et al (2014); Tak et al (2018).…”
Section: Related Workmentioning
confidence: 99%