2018
DOI: 10.4018/ijswis.2018100101
|View full text |Cite
|
Sign up to set email alerts
|

A New LSA and Entropy-Based Approach for Automatic Text Document Summarization

Abstract: Automatic text document summarization is active research area in text mining field. In this article, the authors are proposing two new approaches (three models) for sentence selection, and a new entropy-based summary evaluation criteria. The first approach is based on the algebraic model, Singular Value Decomposition (SVD), i.e. Latent Semantic Analysis (LSA) and model is termed as proposed_model-1, and Second Approach is based on entropy that is further divided into proposed_model-2 and proposed_model-3. In f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 54 publications
0
5
0
1
Order By: Relevance
“…Extractive summarization has been applied to several domains such as automatic highlighting of text (Grootjen & Kachergis, 2018), web articles, multi-document summarization (Gupta & Siddiqui, 2012;Wang et al, 2008) and several others. A number of diversified techniques have been employed for extractive summarization such as statistical-based approaches (Ferreira et al, 2013), genetic algorithm (Sim on et al, 2018, graph-based (Erkan & Radev, 2004;Mihalcea & Tarau, 2004), neural networks (Sinha et al, 2018), optimization-based (Gillick et al, 2008;McDonald, 2007), semantic similarity-based (Gong & Liu, 2001;Steinberger & Ježek, 2004;Yadav & Sharan, 2018), fuzzy-logic based (Abbasi-ghalehtaki et al, 2016;Valladares-vald et al, 2020) and centroid-based techniques (Radev et al, 2004). This work focuses on the previous research on the unsupervised approach.…”
Section: Extractive Summarizationmentioning
confidence: 99%
“…Extractive summarization has been applied to several domains such as automatic highlighting of text (Grootjen & Kachergis, 2018), web articles, multi-document summarization (Gupta & Siddiqui, 2012;Wang et al, 2008) and several others. A number of diversified techniques have been employed for extractive summarization such as statistical-based approaches (Ferreira et al, 2013), genetic algorithm (Sim on et al, 2018, graph-based (Erkan & Radev, 2004;Mihalcea & Tarau, 2004), neural networks (Sinha et al, 2018), optimization-based (Gillick et al, 2008;McDonald, 2007), semantic similarity-based (Gong & Liu, 2001;Steinberger & Ježek, 2004;Yadav & Sharan, 2018), fuzzy-logic based (Abbasi-ghalehtaki et al, 2016;Valladares-vald et al, 2020) and centroid-based techniques (Radev et al, 2004). This work focuses on the previous research on the unsupervised approach.…”
Section: Extractive Summarizationmentioning
confidence: 99%
“…Since high entropy of a sentence implies more coverage, the method has inherent bias towards long sentences, favoring their inclusion in summary. Yadav & Sharan (2018) gauge entropy in latent semantic space of the document by computing probability distribution of topics in sentences and vice-versa. They use Latent Semantic Analysis (LSA) to reveal the latent semantic space of the document.…”
Section: Entropy For Document Summarizationmentioning
confidence: 99%
“…It has been used in several areas of applications such as web articles [2], meeting and telephonic conversations [3], multi-document summarization [1], [38], [39], automatic highlighting of text [40] and many others. To produce an extractive summary various technique are used such as genetic algorithm [41], conditional random fields [42], neural networks [43], semantic similarity such as latent semantic analysis [44]- [46]. Along with these supervised learning techniques, unsupervised learning methods such as fuzzy logic [28], [29], [47], [48], k-means clustering along with term frequency-inverse document frequency [49] has been used.…”
Section: A Extractive Summarizationmentioning
confidence: 99%