Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1510
|View full text |Cite
|
Sign up to set email alerts
|

Encode, Tag, Realize: High-Precision Text Editing

Abstract: We propose LASERTAGGER-a sequence tagging approach that casts text generation as a text editing task. Target texts are reconstructed from the inputs using three main edit operations: keeping a token, deleting it, and adding a phrase before the token. To predict the edit operations, we propose a novel model, which combines a BERT encoder with an autoregressive Transformer decoder. This approach is evaluated on English text on four tasks: sentence fusion, sentence splitting, abstractive summarization, and gramma… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
88
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 116 publications
(102 citation statements)
references
References 43 publications
0
88
0
Order By: Relevance
“…Currently, the only sequence editing model to be applied to GEC is LaserTagger (Malmi et al, 2019). Similarly to the two previously cited works, LaserTagger learns to edit sentences by two different edit operations: KEEP and DELETE, along with pairing these operations with a limited phrase vocabulary consisting of tokens that are frequently changed between the source and target sequences.…”
Section: Sequence Editing Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Currently, the only sequence editing model to be applied to GEC is LaserTagger (Malmi et al, 2019). Similarly to the two previously cited works, LaserTagger learns to edit sentences by two different edit operations: KEEP and DELETE, along with pairing these operations with a limited phrase vocabulary consisting of tokens that are frequently changed between the source and target sequences.…”
Section: Sequence Editing Modelsmentioning
confidence: 99%
“…As described in Malmi et al (2019), a sequence editing model learns to generate a target sentence by applying a small set of edit operations to the source sentence. It works in three steps: (1) the input sentence is encoded into a hidden representation, (2) each token in the input sentence is assigned an edit tag, and (3) rules are applied to convert the output tags into tokens.…”
Section: Sequence Editing Modelmentioning
confidence: 99%
“…Sentence simplification (Nisioi et al, 2017) aims at using techniques such as shortening the sentences to make a text more readable. On the other hand, style transfer is the task of making an utterance conform to a specific style such as formality (Logeswaran et al, 2018;Sennrich et al, 2016).…”
Section: Sentence Editing and Simplificationmentioning
confidence: 99%
“…Tagging solves text editing in two steps instead. It firstly employs a seq2seq framework to produce tag sequences, and secondly, edits input texts according to the tag sequences (the "realization" step) (Malmi et al, 2019). Tagging assigns the tag KEEP for words that do not need to be changed so that it does not need to learn a copy mechanism.…”
Section: Related Workmentioning
confidence: 99%
“…Upon receiving the encoder's hidden states that comprise the source text information, the decoder of End2end directly decodes the hidden states and generates the completely edited target text sequence. But, the decoder of Tagging produces a sequence of editing operations, such as deletion and insertion, that is later applied to the source text to yield the edited text via a realization step (Malmi et al, 2019). The mechanisms of End2end and Tagging are illustrated in Figure 1.…”
Section: Introductionmentioning
confidence: 99%