2020
DOI: 10.48550/arxiv.2009.00713
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

WaveGrad: Estimating Gradients for Waveform Generation

Nanxin Chen,
Yu Zhang,
Heiga Zen
et al.

Abstract: This paper introduces WaveGrad, a conditional model for waveform generation through estimating gradients of the data density. This model is built on the prior work on score matching and diffusion probabilistic models. It starts from Gaussian white noise and iteratively refines the signal via a gradient-based sampler conditioned on the mel-spectrogram. WaveGrad is non-autoregressive, and requires only a constant number of generation steps during inference. It can use as few as 6 iterations to generate high fide… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
108
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 64 publications
(119 citation statements)
references
References 27 publications
0
108
0
Order By: Relevance
“…x 0 = f . Moreover, to make the network G θ aware of the level of noise, we directly give the adequate statistics for the variance of the noise α to the network, similar to [8,34].…”
Section: Proposed Methods 31 Framework Of Diffusemorphmentioning
confidence: 99%
See 2 more Smart Citations
“…x 0 = f . Moreover, to make the network G θ aware of the level of noise, we directly give the adequate statistics for the variance of the noise α to the network, similar to [8,34].…”
Section: Proposed Methods 31 Framework Of Diffusemorphmentioning
confidence: 99%
“…where z ∼ N (0, I). Here, in choosing the total sampling steps T , we employ [8] that presents a more efficient inference method than DDPM. Due to the directly conditioning our denoising model on the noise level α, one can flexibly set the number of sampling steps.…”
Section: Image Registration Using Diffusemorphmentioning
confidence: 99%
See 1 more Smart Citation
“…We show results under trajectories of different number of timesteps K. We select the minimum K such that analytic-DPM can outperform the baselines with full timesteps and underline the corresponding results. (Chen et al, 2020;Kong et al, 2020;Popov et al, 2021;Lam et al, 2021), controllable generation (Choi et al, 2021;Sinha et al, 2021), image super-resolution (Saharia et al, 2021;, image-to-image translation (Sasaki et al, 2021), shape generation (Zhou et al, 2021) and time series forecasting (Rasul et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…It would be interesting to apply Analytic-DPM to other data modalities, e.g. speech data (Chen et al, 2020). As presented in Appendix E, our method can be applied to continuous DPMs, e.g., variational diffusion models (Kingma et al, 2021) that learn the forward noise schedule.…”
Section: H4 Future Workmentioning
confidence: 99%