Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.114
|View full text |Cite
|
Sign up to set email alerts
|

Reflective Decoding: Beyond Unidirectional Generation with Off-the-Shelf Language Models

Abstract: Publicly available, large pretrained Language Models (LMs) generate text with remarkable quality, but only sequentially from left to right. As a result, they are not immediately applicable to generation tasks that break the unidirectional assumption, such as paraphrasing or text-infilling, necessitating task-specific supervision.In this paper, we present REFLECTIVE DE-CODING, a novel unsupervised algorithm that allows for direct application of unidirectional LMs to non-sequential tasks. Our 2-step approach req… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 20 publications
0
4
0
Order By: Relevance
“…Unfortunately, standard left-to-right language models cannot directly infill text, and popular masked language models are mainly trained to infill very short spans [17,22,52,54]. Recent work addresses this by changing model architectures, inference procedures, and training objectives [2,59,67,3]. Most related to our approach is the work of Donahue et al [23] and CM3 [2], who train left-to-right language models to fill in masked token segments of varying lengths; and the work of Alon et al [7], who train an infilling-capable, AST-structured generative model of code on a smaller scale.…”
Section: Related Workmentioning
confidence: 99%
“…Unfortunately, standard left-to-right language models cannot directly infill text, and popular masked language models are mainly trained to infill very short spans [17,22,52,54]. Recent work addresses this by changing model architectures, inference procedures, and training objectives [2,59,67,3]. Most related to our approach is the work of Donahue et al [23] and CM3 [2], who train left-to-right language models to fill in masked token segments of varying lengths; and the work of Alon et al [7], who train an infilling-capable, AST-structured generative model of code on a smaller scale.…”
Section: Related Workmentioning
confidence: 99%
“…In Figure 2 (c), we quantify the exposure bias problem through human judgements. We first sample 200 ELI5 test set questions and generate answers of various lengths {80, 100, ..., 260} (260 is the average sequence length in training set) with beam search, sampling, reflective (West et al, 2021), and KID. We then ask humans to rate these generations with 7-point Likert scoring (Joshi et al, 2015) how likely the generated text is a natural sentence.…”
Section: Eli5 Wowmentioning
confidence: 99%
“…None of these decoding algorithms consider integrating knowledge in the generation process. Reflective decoding (West et al, 2021) and DeLorean (Qin et al, 2020) are two recent decoding algorithms that focus on abductive commonsense reasoning. Reflective decoding in particular has the potential to be extended to other knowledge-intensive tasks.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation