Proceedings of the 2nd Workshop on Neural Machine Translation and Generation 2018
DOI: 10.18653/v1/w18-2706
|View full text |Cite
|
Sign up to set email alerts
|

Controllable Abstractive Summarization

Abstract: Current models for document summarization disregard user preferences such as the desired length, style, the entities that the user might be interested in, or how much of the document the user has already read. We present a neural summarization model with a simple but effective mechanism to enable users to specify these high level attributes in order to control the shape of the final summaries to better suit their needs. With user input, our system can produce high quality summaries that follow user preferences… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
200
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 217 publications
(211 citation statements)
references
References 22 publications
0
200
0
Order By: Relevance
“…Very few works explicitly define a bernoulli distribution for the selector, then train with the REINFORCE algorithm (Ling and Rush, 2017;Chen and Bansal, 2018), but the selection targets at a high recall regardless of the low precision, so the controllability over generated text is weak. Fan et al (2018) control the generation by manually concatenating entity embeddings, while our model is much more flexible by explicitly defining the selection probability over all source tokens.…”
Section: Related Workmentioning
confidence: 99%
“…Very few works explicitly define a bernoulli distribution for the selector, then train with the REINFORCE algorithm (Ling and Rush, 2017;Chen and Bansal, 2018), but the selection targets at a high recall regardless of the low precision, so the controllability over generated text is weak. Fan et al (2018) control the generation by manually concatenating entity embeddings, while our model is much more flexible by explicitly defining the selection probability over all source tokens.…”
Section: Related Workmentioning
confidence: 99%
“…These systems can be used in various application areas, such as text summarization (Fan et al, 2018), adversarial example generation (Iyyer et al, 2018), dialogue (Niu and Bansal, 2018), and data-to-document generation (Wiseman et al, 2018). However, prior work on controlled generation has typically assumed a known, finite set of values that the controlled attribute can take on.…”
Section: Introductionmentioning
confidence: 99%
“…To investigate whether the proposed method can generate good headlines for unseen lengths, we excluded headlines whose lengths are equal to the desired length (len) from the training data. The len = 10 len = 13 len = 26 32.85 11.78 28.52 Previous studies for controlling output length Kikuchi et al (2016) 26.73 8.39 23.88 Fan et al (2018) 30.00 10.27 26.43 Other previous studies Rush et al (2015) 28.18 8.49 23.81 Nagata (2017) 32.28 10.54 27.80 Zhou et al (2017) 29.21 9.56 25.51 Li et al (2017) 31.79 10.75 27.48 Li et al (2018) 29.33 10.24 25.24 Table 3 shows the recall-oriented ROUGE scores on the DUC-2004 test set. Following the evaluation protocol (Over et al, 2007), we truncated characters over 75 bytes.…”
Section: Resultsmentioning
confidence: 99%
“…In fact, Figure 1 shows a large variance in output sequences produced by a widely used encoder-decoder model (Luong et al, 2015), which has no mechanism for controlling the length of the output sequences. Fan et al (2018) trained embeddings that correspond to each output length to control the output sequence length. Since the embeddings for different lengths are independent, it is hard to generate a sequence of the length that is infrequent in training data.…”
Section: Introductionmentioning
confidence: 99%