Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-1158
|View full text |Cite
|
Sign up to set email alerts
|

Ranking Sentences for Extractive Summarization with Reinforcement Learning

Abstract: Single document summarization is the task of producing a shorter version of a document while preserving its principal information content. In this paper we conceptualize extractive summarization as a sentence ranking task and propose a novel training algorithm which globally optimizes the ROUGE evaluation metric through a reinforcement learning objective. We use our algorithm to train a neural summarization model on the CNN and Dai-lyMail datasets and demonstrate experimentally that it outperforms state-of-the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
442
0
4

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 469 publications
(470 citation statements)
references
References 46 publications
4
442
0
4
Order By: Relevance
“…The fourth block reports results with fine-tuned BERT models: BERTSUMEXT and its two variants (one without interval embeddings, and one with the large version of BERT), BERTSUM-ABS, and BERTSUMEXTABS. BERT-based models outperform the LEAD-3 baseline which is not a strawman; on the CNN/DailyMail corpus it is indeed superior to several extractive (Nallapati et al, 2017;Narayan et al, 2018b; and abstractive models (See et al, 2017). BERT models collectively outperform all previously proposed extractive and abstractive systems, only falling behind the ORACLE upper bound.…”
Section: Automatic Evaluationmentioning
confidence: 82%
“…The fourth block reports results with fine-tuned BERT models: BERTSUMEXT and its two variants (one without interval embeddings, and one with the large version of BERT), BERTSUM-ABS, and BERTSUMEXTABS. BERT-based models outperform the LEAD-3 baseline which is not a strawman; on the CNN/DailyMail corpus it is indeed superior to several extractive (Nallapati et al, 2017;Narayan et al, 2018b; and abstractive models (See et al, 2017). BERT models collectively outperform all previously proposed extractive and abstractive systems, only falling behind the ORACLE upper bound.…”
Section: Automatic Evaluationmentioning
confidence: 82%
“…On top of the seq2seq framework, many other variants have been studied using convolutional networks (Cheng and Lapata, 2016;Allamanis et al, 2016), pointer networks (See et al, 2017), scheduled sampling (Bengio et al, 2015), and reinforcement learning (Paulus et al, 2017). In extractive systems, different types of encoders (Cheng and Lapata, 2016;Nallapati et al, 2017;Kedzie et al, 2018) and optimization techniques (Narayan et al, 2018b) have been developed. Our goal is to explore which types of systems learns which sub-aspect of summarization.…”
Section: Related Workmentioning
confidence: 99%
“…Extractive: [14] use hierarchical Recurrent Neural Networks (RNNs) to get the representations of the sentences and classify the importance of sentences. [15] rank extracted sentences for summary generation through a reinforcement learning and [16] extract salient sentences and propose a new policy gradient method to rewrite these sentences (i.e., compresses and paraphrases) to generate a concise overall summary. [17] propose a framework composed of a hierarchical document encoder based on CNNs and an attention-based extractor with attention over external information.…”
Section: Related Workmentioning
confidence: 99%