2018
DOI: 10.48550/arxiv.1807.11143
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ARM: Augment-REINFORCE-Merge Gradient for Stochastic Binary Networks

Mingzhang Yin,
Mingyuan Zhou

Abstract: To backpropagate the gradients through discrete stochastic layers, we encode the true gradients into a multiplication between random noises and the difference of the same function of two different sets of discrete latent variables, which are correlated with these random noises. The expectations of that multiplication over iterations are zeros combined with spikes from time to time. To modulate the frequencies, amplitudes, and signs of the spikes to capture the temporal evolution of the true gradients, we propo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(11 citation statements)
references
References 6 publications
(10 reference statements)
0
11
0
Order By: Relevance
“…Linderman et al (2018) relaxes the discrete set of permutations to Birkhoff polytope, the set of doubly-stochastic matrices, and extend stick-breaking approach (Sethuraman 2017) and derive control variate for black-box function optimization combining the REINFORCE estimator and reparametrization trick. Yin and Zhou (2018) propose gradient estimator that estimates the gradients of discrete distribution parameters in an augmented space.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Linderman et al (2018) relaxes the discrete set of permutations to Birkhoff polytope, the set of doubly-stochastic matrices, and extend stick-breaking approach (Sethuraman 2017) and derive control variate for black-box function optimization combining the REINFORCE estimator and reparametrization trick. Yin and Zhou (2018) propose gradient estimator that estimates the gradients of discrete distribution parameters in an augmented space.…”
Section: Related Workmentioning
confidence: 99%
“…Despite the recent breakthroughs in gradient estimation for continuous latent variables (Kingma and Welling 2013;Rezende, Mohamed, and Wierstra 2014;Mohamed et al 2019), gradient estimation for discrete latent variables remains a challenge. Currently, general-purpose estimators (Williams 1992;Mnih and Gregor 2014) remain unreliable and the state-of-the-art methods (Tucker et al 2017;Grathwohl et al 2018;Yin and Zhou 2018) exclusively consider the categorical distribution. Although the reduction to the categorical case allows benefiting from gradient estimators for continuous relaxations, such solutions are hard to translate to discrete distributions with large support.…”
Section: Introductionmentioning
confidence: 99%
“…Attracted by the generalization of REINFORCE, many works try to improve its performance via efficient variance-reduction techniques, like control variants (Mnih & Gregor, 2014;Titsias & Lázaro-Gredilla, 2015;Gu et al, 2015;Mnih & Rezende, 2016;Tucker et al, 2017;Grathwohl et al, 2017) or via data augmentation and permutation techniques (Yin & Zhou, 2018). Most of this research focuses on discrete random variables, likely because Rep (if it exists) works well for continuous random variables but it may not exist for discrete random variables.…”
Section: Related Workmentioning
confidence: 99%
“…SF suffers from high variance and many remedies have been proposed to reduce this variance (Mnih & Gregor (2014); Gregor et al (2013); Gu et al (2015); Mnih & Rezende (2016); Tucker et al (2017); Grathwohl et al (2017)). Unbiased estimators that require multiple function evaluations have also been proposed Tokui & Sato (2017), Titsias & Lázaro-Gredilla (2015), Lorberbom et al (2018), Yin & Zhou (2018). These estimators have lower variance but can be computationally demanding.…”
Section: Related Workmentioning
confidence: 99%
“…FD estimators are computationally expensive and require multiple function evaluations per gradient. A notable exception is the Augment-REINFORCE-Merge (ARM) estimator introduced in Yin & Zhou (2018). ARM provides an unbiased estimate using only two function evaluations for the factorized multivariate distribution, regardless of the number of variables.…”
Section: The Augment-reinforce-merge Estimatormentioning
confidence: 99%