Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2014
DOI: 10.3115/v1/d14-1075
|View full text |Cite
|
Sign up to set email alerts
|

Fear the REAPER: A System for Automatic Multi-Document Summarization with Reinforcement Learning

Abstract: This paper explores alternate algorithms, reward functions and feature sets for performing multi-document summarization using reinforcement learning with a high focus on reproducibility. We show that ROUGE results can be improved using a unigram and bigram similarity metric when training a learner to select sentences for summarization. Learners are trained to summarize document clusters based on various algorithms and reward functions and then evaluated using ROUGE. Our experiments show a statistically signifi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
34
0
1

Year Published

2015
2015
2020
2020

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 30 publications
(36 citation statements)
references
References 22 publications
1
34
0
1
Order By: Relevance
“…The use of reinforcement learning (RL) in extractive summarization was first explored by Ryang and Abekawa (2012), who proposed to use the TD(λ) algorithm to learn a value function for sentence selection. Rioux et al (2014) improved this framework by replacing the learning agent with another TD(λ) algorithm. However, the performance of their methods was limited by the use of shallow function approximators, which required performing a fresh round of reinforcement learning for every new document to be summarized.…”
Section: Related Workmentioning
confidence: 99%
“…The use of reinforcement learning (RL) in extractive summarization was first explored by Ryang and Abekawa (2012), who proposed to use the TD(λ) algorithm to learn a value function for sentence selection. Rioux et al (2014) improved this framework by replacing the learning agent with another TD(λ) algorithm. However, the performance of their methods was limited by the use of shallow function approximators, which required performing a fresh round of reinforcement learning for every new document to be summarized.…”
Section: Related Workmentioning
confidence: 99%
“…(8)) α = 10 −3 learning rate for preference learning φ(y, x) vectorised representation of summary y for document cluster x (see Eq. (8)); we use the same vector representation as Rioux et al (2014)…”
Section: Parameter Descriptionmentioning
confidence: 99%
“…[0,10]). For the vector representation φ, we use the same 200-dimensional bag-of-bigram representation as Rioux et al (2014) (w d = 1) .288 * .297 * .319 * den (we = 1)…”
Section: Parameter Descriptionmentioning
confidence: 99%
See 1 more Smart Citation
“…Many works have viewed the summarization problem as a supervised classification problem in which several features are used to predict the inclusion of document sentences in the summary. Variations of supervised models have been utilized for summary generation, such as: maximum entropy (Osborne, 2002), HMM (Conroy et al, 2011), CRF (Galley, 2006;Shen et al, 2007;Chali and Hasan, 2012), SVM (Xie and Liu, 2010), logistic regression (Louis et al, 2010) and reinforcement learning (Rioux et al, 2014). Problems with supervised models in context of summarization include the need for large amount of annotated data and domain dependency.…”
Section: Reference Articlementioning
confidence: 99%