Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics 2019
DOI: 10.18653/v1/p19-1101
|View full text |Cite
|
Sign up to set email alerts
|

A Simple Theoretical Model of Importance for Summarization

Abstract: Research on summarization has mainly been driven by empirical approaches, crafting systems to perform well on standard datasets with the notion of information Importance remaining latent. We argue that establishing theoretical models of Importance will advance our understanding of the task and help to further improve summarization systems. To this end, we propose simple but rigorous definitions of several concepts that were previously used only intuitively in summarization: Redundancy, Relevance, and Informati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
65
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 69 publications
(70 citation statements)
references
References 60 publications
0
65
0
Order By: Relevance
“…Our study extends the hypothesis to various corpora as well as systems. With a specific focus on importance aspect, a recent work (Peyrard, 2019a) divided it into three subcategories; redundancy, relevance, and informativeness, and provided quantities of each to measure. Compared to this, ours provide broader scale of sub-aspect analysis across various corpora and systems.…”
Section: Related Workmentioning
confidence: 99%
“…Our study extends the hypothesis to various corpora as well as systems. With a specific focus on importance aspect, a recent work (Peyrard, 2019a) divided it into three subcategories; redundancy, relevance, and informativeness, and provided quantities of each to measure. Compared to this, ours provide broader scale of sub-aspect analysis across various corpora and systems.…”
Section: Related Workmentioning
confidence: 99%
“…This criterion casts summarization as finding a set of summary sentences which closely match the doc distribution. When selecting sentences to constitute the summary, this optimization objective penalizes redundancy while maximizing relevance [17]. Because the problem of finding the subset of sentences from a collection that minimizes the KL divergence is NP-hard, a greedy algorithm is often used in practice.…”
Section: B Bayesian Approaches In Summarizationmentioning
confidence: 99%
“…Despite its fundamental role, background knowledge has received little attention from the summarization community. Existing approaches largely focus on the relevance aspect, which enforces similarity between the generated summaries and the source documents (Peyrard, 2019). Figure 1: A summary (S) results from the combination of the background knowledge (K) and the source document (D).…”
Section: Introductionmentioning
confidence: 99%
“…Figure 1: A summary (S) results from the combination of the background knowledge (K) and the source document (D). Following Peyrard (2019), S is similar to D (Relevance measured by a small KL(S||D)) but also brings new information compared to background knowledge (informativeness measured by a large KL(S||K)). We can infer the unobserved K from the choices unexplained by the Relevance criteria.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation