2021 5th International Conference on Trends in Electronics and Informatics (ICOEI) 2021
DOI: 10.1109/icoei51242.2021.9453071
|View full text |Cite
|
Sign up to set email alerts
|

Text Summarization using TF-IDF and Textrank algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 18 publications
0
6
0
Order By: Relevance
“…TextRank is an unsupervised method that has been used in many previous work on summarization, either as main or baseline method, because of its proven effectiveness [8]- [11]. Previously, many studies have also been conducted to enhance the performance of the TextRank algorithm for text summarization or keyword extraction, such as using word embedding [12]- [18], term frequency-inverse document frequency (TF-IDF) [19], [20], the combination of 1 gram, 2 gram, and Hidden Markov models [21], knowledge graph sentence embedding and K-means clustering [22], statistical and linguistic features for sentence weighting [23], variation of sentence similarity functions [24], and fine-tuning the hyperparameters [13].…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…TextRank is an unsupervised method that has been used in many previous work on summarization, either as main or baseline method, because of its proven effectiveness [8]- [11]. Previously, many studies have also been conducted to enhance the performance of the TextRank algorithm for text summarization or keyword extraction, such as using word embedding [12]- [18], term frequency-inverse document frequency (TF-IDF) [19], [20], the combination of 1 gram, 2 gram, and Hidden Markov models [21], knowledge graph sentence embedding and K-means clustering [22], statistical and linguistic features for sentence weighting [23], variation of sentence similarity functions [24], and fine-tuning the hyperparameters [13].…”
Section: Introductionmentioning
confidence: 99%
“…Then, TF-IDF is used to weigh the generated word embedding to further improve sentence representation; this is based on our intuition that more important words, which are estimated using corpus statistics, should be valued more when generating the vector representation of sentences. TF-IDF is chosen because it has been shown to perform well for term weighting in various natural language processing tasks [19], [20], [25]- [27], and it also has been proven to significantly outperform the bag-of-words (BoW) technique [28]. The combination of word embedding and TF-IDF weighting components are then expected to improve the estimation of sentence relationships in the TextRank algorithm.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations