2019
DOI: 10.48550/arxiv.1906.10652
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Monte Carlo Gradient Estimation in Machine Learning

Abstract: This paper is a broad and accessible survey of the methods we have at our disposal for Monte Carlo gradient estimation in machine learning and across the statistical sciences: the problem of computing the gradient of an expectation of a function with respect to parameters defining the distribution that is integrated; the problem of sensitivity analysis. In machine learning research, this gradient problem lies at the core of many learning problems, in supervised, unsupervised and reinforcement learning. We will… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
51
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(51 citation statements)
references
References 69 publications
(99 reference statements)
0
51
0
Order By: Relevance
“…In fact, when the ansatz is close to an eigenstate of Ĥ, then E loc (σ) ≈ E, which means that the variance of gradients Var(∂ λj E) ≈ 0 for each variational parameter λ j . We note that this is similar in spirit to the control variate methods in Monte Carlo and to the baseline methods in reinforcement learning [51].…”
Section: Variational Monte Carlomentioning
confidence: 59%
See 1 more Smart Citation
“…In fact, when the ansatz is close to an eigenstate of Ĥ, then E loc (σ) ≈ E, which means that the variance of gradients Var(∂ λj E) ≈ 0 for each variational parameter λ j . We note that this is similar in spirit to the control variate methods in Monte Carlo and to the baseline methods in reinforcement learning [51].…”
Section: Variational Monte Carlomentioning
confidence: 59%
“…Here, we can subtract the term E in order to reduce noise in the stochastic estimation of our gradients without introducing a bias [20,51]. In fact, when the ansatz is close to an eigenstate of Ĥ, then E loc (σ) ≈ E, which means that the variance of gradients Var(∂ λj E) ≈ 0 for each variational parameter λ j .…”
Section: Variational Monte Carlomentioning
confidence: 99%
“…We remark that π in the agent optimization (7a) and π in the adversary optimization (7b) are different, as are the pessimistic and optimistic hallucination policies η (p) and η (o) . In particular, we learn a critic via fitted Q-iteration (Perolat et al, 2015;Antos et al, 2008) and then differentiate through the critic using pathwise gradients (Mohamed et al, 2019;Silver et al, 2014) using stochastic gradient ascent for the agent and stochastic gradient descent for the adversary.…”
Section: Practical Implementationmentioning
confidence: 99%
“…This improves training stability and is a standard technique deployed in machine learning to improve gradient estimation [38,39].…”
Section: Training Detailsmentioning
confidence: 99%