Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics 2014
DOI: 10.3115/v1/e14-1075
|View full text |Cite
|
Sign up to set email alerts
|

Improving the Estimation of Word Importance for News Multi-Document Summarization

Abstract: We introduce a supervised model for predicting word importance that incorporates a rich set of features. Our model is superior to prior approaches for identifying words used in human summaries.Moreover we show that an extractive summarizer using these estimates of word importance is comparable in automatic evaluation with the state-of-the-art.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
83
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 105 publications
(90 citation statements)
references
References 30 publications
0
83
0
Order By: Relevance
“…We also choose as baselines those state-of-the-art summarization results on DUC (2001, 2002, and 2004) data. To our knowledge, the best reported results on DUC 2001DUC , 2002DUC and 2004 are from R2N2 (Cao et al, 2015), ClusterCMRW (Wan and Yang, 2008) and REG-SUM 2 (Hong and Nenkova, 2014) respectively. R2N2 applies recursive neural networks to learn 2 REGSUM truncates a summary to 100 words.…”
Section: Comparison With Baseline Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We also choose as baselines those state-of-the-art summarization results on DUC (2001, 2002, and 2004) data. To our knowledge, the best reported results on DUC 2001DUC , 2002DUC and 2004 are from R2N2 (Cao et al, 2015), ClusterCMRW (Wan and Yang, 2008) and REG-SUM 2 (Hong and Nenkova, 2014) respectively. R2N2 applies recursive neural networks to learn 2 REGSUM truncates a summary to 100 words.…”
Section: Comparison With Baseline Methodsmentioning
confidence: 99%
“…In previous summarization systems, though not well-studied, some widely-used sentence ranking features such as the length and the ratio of stopwords, can be seen as attempts to measure the summary prior nature to a certain extent. Notably, Hong and Nenkova (2014) built a state-of-the-art summarization system through making use of advanced document-independent features. However, these document-independent features are usually hand-crafted, difficult to exhaust each aspect of the summary prior nature.…”
Section: Introductionmentioning
confidence: 99%
“…Although this dataset has mainly been used to train extractive summarization systems (Hong and Nenkova, 2014;Hong et al, 2015;Li et al, 2016;Durrett et al, 2016), it has recently been used for the abstractive summarization task (Paulus et al, 2018). NYT dataset (Sandhaus, 2008) is a collection of articles published between 1996 and 2007.…”
Section: New York Times (Nyt)mentioning
confidence: 99%
“…As TLS organizes events by date, timelines can be generated by MDS systems (such as (Radev et al, 2004b;Radev et al, 2004a;McKeown et al, 2003;Erkan and Radev, 2004;Metzler and Kanungo, 2008;Hong and Nenkova, 2014) by applying their summarization techniques on news articles for every individual date to create corresponding daily summaries. However, manually written timelines normally only include a small number of dates; in addition, the temporal component imposes constraints on sentence selection for timeline summarization, such as the preference for little overlap between sentences selected for different dates (Yan et al, 2011b).…”
Section: Related Workmentioning
confidence: 99%