The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2019
DOI: 10.1162/tacl_a_00292
|View full text |Cite
|
Sign up to set email alerts
|

Insertion-based Decoding with Automatically Inferred Generation Order

Abstract: Conventional neural autoregressive decoding commonly assumes a fixed left-to-right generation order, which may be sub-optimal. In this work, we propose a novel decoding algorithm -InDIGO -which supports flexible sequence generation in arbitrary orders through insertion operations. We extend Transformer, a state-of-the-art sequence generation model, to efficiently implement the proposed approach, enabling it to be trained with either a pre-defined generation order or adaptive orders obtained from beam-search. E… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
82
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 85 publications
(86 citation statements)
references
References 30 publications
0
82
0
Order By: Relevance
“…We train with a simple masking scheme where the number of masked target tokens is distributed uniformly, presenting the model with both easy (single mask) and difficult (completely masked) examples. Unlike recently proposed insertion models (Gu et al, 2019;Stern et al, 2019), which treat each token as a separate training instance, CMLMs can train from the entire sequence in parallel, resulting in much faster training.…”
Section: Introductionmentioning
confidence: 99%
“…We train with a simple masking scheme where the number of masked target tokens is distributed uniformly, presenting the model with both easy (single mask) and difficult (completely masked) examples. Unlike recently proposed insertion models (Gu et al, 2019;Stern et al, 2019), which treat each token as a separate training instance, CMLMs can train from the entire sequence in parallel, resulting in much faster training.…”
Section: Introductionmentioning
confidence: 99%
“…In parallel to the work investigating masked language models for text generation, Welleck et al [74], Stern et al [75] and Gu et al [76] proposed methods for non-monotonic sequential text generation. Although these methods could be applied for generating molecular graphs in flexible ordering, there has not been work empirically validating this.…”
Section: A Related Workmentioning
confidence: 99%
“…In recent work, several insertion-based frameworks have been proposed for the generation of sequences in a non-left-to-right fashion for machine translation (Stern et al, 2019;Welleck et al, 2019;Gu et al, 2019). Stern et al (2019) and balanced binary tree orders.…”
Section: Related Workmentioning
confidence: 99%
“…More recently, a number of novel insertionbased architectures have been developed for sequence generation (Gu et al, 2019;Stern et al, 2019;Welleck et al, 2019). These frameworks license a diverse set of generation orders, including uniform (Welleck et al, 2019), random (Gu et al, 2019), or balanced binary trees (Stern et al, 2019). Some of them also match the quality of state-ofthe-art left-to-right models (Stern et al, 2019).…”
Section: Introductionmentioning
confidence: 99%