Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1140
|View full text |Cite
|
Sign up to set email alerts
|

Controlling Output Length in Neural Encoder-Decoders

Abstract: Neural encoder-decoder models have shown great success in many sequence generation tasks. However, previous work has not investigated situations in which we would like to control the length of encoder-decoder outputs. This capability is crucial for applications such as text summarization, in which we have to generate concise summaries with a desired length. In this paper, we propose methods for controlling the output sequence length for neural encoder-decoder models: two decoding-based methods and two learning… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
164
0
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 187 publications
(175 citation statements)
references
References 28 publications
0
164
0
1
Order By: Relevance
“…With lstm+leninit, 2 out of 100 sentences, chosen randomly, ended with a word that cannot be located at the end of a sentence. In contrast, with lstm+lenemb, 24 sentences ended with such words and therefore are ungrammatical, although lenemb has shown to be effective in abstractive sentence summarization (Kikuchi et al, 2016). This result suggests that lstm+lenemb is excessively affected by the desired length because lenemb receives the potential desired length at each time of decoding.…”
Section: Human Evaluationmentioning
confidence: 95%
See 1 more Smart Citation
“…With lstm+leninit, 2 out of 100 sentences, chosen randomly, ended with a word that cannot be located at the end of a sentence. In contrast, with lstm+lenemb, 24 sentences ended with such words and therefore are ungrammatical, although lenemb has shown to be effective in abstractive sentence summarization (Kikuchi et al, 2016). This result suggests that lstm+lenemb is excessively affected by the desired length because lenemb receives the potential desired length at each time of decoding.…”
Section: Human Evaluationmentioning
confidence: 95%
“…As the second and the third models, we extend the first model to control the output length (Kikuchi et al, 2016). The second model, lstm+leninit, initializes the memory cell of the decoder as follows: m 0 = tarlen * b len where tarlen is the desired output length, and b len is a trainable parameter.…”
Section: Sentence Compression With Lstmmentioning
confidence: 99%
“…Each control variable corresponds to an attribute of a sentence. Compared to previous work that only seeks to control the values of sentiment (Hu et al, 2017) and length (Kikuchi et al, 2016), we further explore neural text generation with particular verbal predicates, semantic frames, and automatically-induced clusters.…”
Section: Introductionmentioning
confidence: 99%
“…Thus, we expect to decisively prohibit excessive generation. Finally, we evaluate the effectiveness of our method on well-studied ABS benchmark data provided by Rush et al (2015), and evaluated in (Chopra et al, 2016;Nallapati et al, 2016b;Kikuchi et al, 2016;Takase et al, 2016;Ayana et al, 2016;Gulcehre et al, 2016).…”
Section: Introductionmentioning
confidence: 99%