2022
DOI: 10.48550/arxiv.2210.08933
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models

Abstract: Recently, diffusion models have emerged as a new paradigm for generative models. Despite the success in domains using continuous signals such as vision and audio, adapting diffusion models to natural language is difficult due to the discrete nature of text. We tackle this challenge by proposing DIFFUSEQ: a diffusion model designed for sequence-to-sequence (SEQ2SEQ) text generation tasks. Upon extensive evaluation over a wide range of SEQ2SEQ tasks, we find DIF-FUSEQ achieving comparable or even better performa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
22
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(57 citation statements)
references
References 15 publications
0
22
0
Order By: Relevance
“…2, DINOISER surpasses or closely approaches CMLM even when MBR=1, while the vanilla DiffusionLM heavily relies on a large number of candidates used for MBR decoding. Besides, DINOISER necessitates much fewer NFEs to achieve strong performance, e.g., only 20 steps, resulting in only 1% to 10% computational consumption and latency compared to previous works (Gong et al, 2022;Dieleman et al, 2022). This manifests that DINOISER is more accurate yet efficient compared to previous diffusion-based sequence learning models.…”
Section: Resultsmentioning
confidence: 95%
See 4 more Smart Citations
“…2, DINOISER surpasses or closely approaches CMLM even when MBR=1, while the vanilla DiffusionLM heavily relies on a large number of candidates used for MBR decoding. Besides, DINOISER necessitates much fewer NFEs to achieve strong performance, e.g., only 20 steps, resulting in only 1% to 10% computational consumption and latency compared to previous works (Gong et al, 2022;Dieleman et al, 2022). This manifests that DINOISER is more accurate yet efficient compared to previous diffusion-based sequence learning models.…”
Section: Resultsmentioning
confidence: 95%
“…We consider IWSLT14 DE↔EN (160K pairs), WMT14 EN↔DE (4.0M pairs), and WMT14 EN↔RO (610K pairs), six machine translation tasks with variant sizes of training data. Additionally, we experiment on two of the datasets introduced by Dif-fuSeq (Gong et al, 2022), including Wiki (Jiang et al, 2020) for text simplification and QQP 6 for paraphrasing.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations