2004
DOI: 10.1007/978-3-540-24618-3_25
|View full text |Cite
|
Sign up to set email alerts
|

Multi-document Automatic Text Summarization Using Entropy Estimates

Abstract: Abstract. This paper describes a sentence ranking technique using entropy measures, in a multi-document unstructured text summarization application. The method is topic specific and makes use of a simple language independent training framework to calculate entropies of symbol units. The document set is summarized by assigning entropy-based scores to a reduced set of sentences obtained using a graph representation for sentence similarity. The performance is seen to be better than some of the common statistical … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2005
2005
2022
2022

Publication Types

Select...
2
2
2

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 6 publications
0
4
0
Order By: Relevance
“…On the other hand, although the concept of entropy and its related derivatives have been successfully applied to a wide range of problems related to summary and compression [19][20][21][22][23], our work is the first attempt to try to involve the computation and control of entropy measurements with the user requirement specification during approximated provenance summarization. We believe that by allowing more flexible control of approximated data provenance summarization using entropy measurements, a wide scope of provenancerelated tasks, for example, provenance based access control rules retrieval [24], provenance visualization [25], and provenance storage [26,27], can be performed both more effectively and more efficiently.…”
Section: Related Workmentioning
confidence: 99%
“…On the other hand, although the concept of entropy and its related derivatives have been successfully applied to a wide range of problems related to summary and compression [19][20][21][22][23], our work is the first attempt to try to involve the computation and control of entropy measurements with the user requirement specification during approximated provenance summarization. We believe that by allowing more flexible control of approximated data provenance summarization using entropy measurements, a wide scope of provenancerelated tasks, for example, provenance based access control rules retrieval [24], provenance visualization [25], and provenance storage [26,27], can be performed both more effectively and more efficiently.…”
Section: Related Workmentioning
confidence: 99%
“…The previous information-theoretic methods for document summarization are based on information distance [17] and entropy estimates [24]. [24] ranks sentences by calculating entropies of symbol units.…”
Section: Related Workmentioning
confidence: 99%
“…The previous information-theoretic methods for document summarization are based on information distance [17] and entropy estimates [24]. [24] ranks sentences by calculating entropies of symbol units. [17] uses the theory of Kolmogorov complexity and generates summaries that have the minimum information distance with the original documents.…”
Section: Related Workmentioning
confidence: 99%
“…Therefore, EnDSUM can generate the summary automatically without explicitly identifying the category of a tweet. Although there are few single and multiple document summarization approaches [1,11,17,19,21] that have highlighted the relevance of entropy based selection of sentences into summary, those approaches are not directly applicable to disaster tweets. The reason being the informal structure of tweets, absence of storyline in tweets and the high vocabulary diversity in user generated tweets.…”
Section: Introductionmentioning
confidence: 99%