2020
DOI: 10.48550/arxiv.2003.01847
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generalized Gumbel-Softmax Gradient Estimator for Generic Discrete Random Variables

Abstract: Estimating the gradients of stochastic nodes is one of the crucial research questions in the deep generative modeling community. This estimation problem becomes further complex when we regard the stochastic nodes to be discrete because pathwise derivative techniques can not be applied. Hence, the gradient estimation requires the score function methods or the continuous relaxation of the discrete random variables. This paper proposes a general version of the Gumbel-Softmax estimator with continuous relaxation, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 4 publications
0
5
0
Order By: Relevance
“…The activation function used is LeakyReLU. Gumbel Softmax outperforms softmax in terms of discrete text production ( Joo et al, 2020 ). After experimenting with various sequence lengths, it was discovered that sequence length 160 produced the best results.…”
Section: Methodsmentioning
confidence: 99%
“…The activation function used is LeakyReLU. Gumbel Softmax outperforms softmax in terms of discrete text production ( Joo et al, 2020 ). After experimenting with various sequence lengths, it was discovered that sequence length 160 produced the best results.…”
Section: Methodsmentioning
confidence: 99%
“…However, directly sampling word counts from the Poisson distribution is not differentiable. In order to enable back-propagation of gradients, we apply Gumbel-Softmax (Jang et al, 2017;Joo et al, 2020), which is a gradient estimator with the reparameterization trick.…”
Section: Distant Supervision and Adversarial Learningmentioning
confidence: 99%
“…Since score methods deviate from the orthodox backpropagation procedure [35], we opted for the latter. Recently, Joo et al [16] introduced the Generalized Gumbel-Softmax (GenGS) reparametrization, which can approximate any discrete, non-negative, and finite-mean random variable. They achieve this by truncating its support to a finite number of bins and relaxing the resulting categorical distribution into a continuous form using the Gumbel-Softmax reparametrization [13].…”
Section: Numerical Approximationmentioning
confidence: 99%