2018
DOI: 10.48550/arxiv.1808.10792
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Bottom-Up Abstractive Summarization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
53
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(53 citation statements)
references
References 0 publications
0
53
0
Order By: Relevance
“…The 'OpenNMT BRNN (2 layer, emb 256, hid 1024)' pre-trained model 4 has been used. • CopyTransformer (Gehrmann et al, 2018) 5 .…”
Section: Abstractive Summarisation Methodologymentioning
confidence: 99%
See 1 more Smart Citation
“…The 'OpenNMT BRNN (2 layer, emb 256, hid 1024)' pre-trained model 4 has been used. • CopyTransformer (Gehrmann et al, 2018) 5 .…”
Section: Abstractive Summarisation Methodologymentioning
confidence: 99%
“…We have carried out an evaluation with 6 abstractive summarisation models: BART (Lewis et al, 2019), T5 (Raffel et al, 2019), BERT (PreSumm -BertSumExtAbs: Liu and Lapata, 2019), PG (Pointer-Generator with Coverage Penalty) (See et al, 2017), CopyTransformer (Gehrmann et al, 2018), and FastAbsRL (Chen and Bansal, 2018). Those models are applied in combination with the machine translation system MarianMT (Junczys-Dowmunt et al, 2018) using the Opus-MT models (Tiedemann and Thottingal, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…Very recently, GEDI (Krause et al, 2020) achieves strong performance by using CCLM generators as discriminators, though it relies on several heuristics. More broadly, text generation models for style transfer (Hu et al, 2017;Lample et al, 2018b;Dai et al, 2019a), summarization (See et al, 2017;Gehrmann et al, 2018;Zaheer et al, 2020), and machine translation (Lample et al, 2018a;Ng et al, 2019;Lewis et al, 2019) can also be viewed as CCLM's for different "attributes. "…”
Section: Related Workmentioning
confidence: 99%
“…Ptr-Net model also adds coverage loss, which examines the difference between the attentions of previous words generated and the current attention, in an attempt to fix the issue of word repetition, a persistent issue in seq2seq models. Gehrmann et al [7] try to improve the fluency of the generated text through various constraints applied during model training. Soft constraints on the size of text are used to constrain the length of generated descriptions, while constraints on the output probability distribution of words ameliorates word repetition.…”
Section: Related Workmentioning
confidence: 99%