Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.420
|View full text |Cite
|
Sign up to set email alerts
|

Blank Language Models

Abstract: We propose Blank Language Model (BLM), a model that generates sequences by dynamically creating and filling in blanks. The blanks control which part of the sequence to expand, making BLM ideal for a variety of text editing and rewriting tasks. The model can start from a single blank or partially completed text with blanks at specified locations. It iteratively determines which word to place in a blank and whether to insert new blanks, and stops generating when no blanks are left to fill. BLM can be efficiently… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
44
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(50 citation statements)
references
References 39 publications
1
44
0
Order By: Relevance
“…Moreover, our padded MLM determines the number of tokens to insert without having to pre-specify it. In that sense, it is similar to the recently proposed Blank Language Model (Shen et al, 2020).…”
Section: Related Worksupporting
confidence: 64%
See 1 more Smart Citation
“…Moreover, our padded MLM determines the number of tokens to insert without having to pre-specify it. In that sense, it is similar to the recently proposed Blank Language Model (Shen et al, 2020).…”
Section: Related Worksupporting
confidence: 64%
“…model, which yields an accuracy of 98.4% on the development set (slightly higher than the CNN classifier used by Shen et al (2020) which has an accuracy of 97.7%). The Exact scores reported in the paper were computed after lowercasing the predictions and the targets.…”
Section: B Hyperparameter Settingsmentioning
confidence: 86%
“…Rudinger et al (2015) frame narrative cloze as a generation task and employ language models, but they only consider one infill of a fixed length. Zhu et al (2019); Shen et al (2020) infill multiple variable-length sequences, but these approaches require the masked context to be iteratively updated and reprocessed to fill in blanks one a time. In contrast, our approach appends infilled text to the context and does not require reprocessing the entire input sequence for each blank.…”
Section: Related Workmentioning
confidence: 99%
“…Previous approaches have proposed alternatives to autoregressive decoding (Gu et al, 2018;Lee et al, 2018;Wang and Cho, 2019). Instead of the left-to-right autoregressive decoding, Insertion Transformer and BLM (Shen et al, 2020) generate the output sequence through insertion operations, whereas LEVT (Gu et al, 2019) additionally incorporates a deletion operation. These methods produce the output iteratively, while FELIX requires only two steps: tagging and insertion.…”
Section: Related Workmentioning
confidence: 99%