2019
DOI: 10.1007/s13369-019-03920-9
|View full text |Cite
|
Sign up to set email alerts
|

On Term Frequency Factor in Supervised Term Weighting Schemes for Text Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 26 publications
(18 citation statements)
references
References 40 publications
0
18
0
Order By: Relevance
“…Some variants of the classical TF scheme are inverse term frequency (ITF) [19], which normalizes the values to the interval [0,1] based on Zipf's Law, and other transformations on the term-frequency values where terms that are extremely frequent do not increase at the same rate as in TF [5]. In [10], Global factors are designed to improve precision although this might be at the expense of a drop in recall. The rationale behind these factors is that common terms are poor discriminators.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Some variants of the classical TF scheme are inverse term frequency (ITF) [19], which normalizes the values to the interval [0,1] based on Zipf's Law, and other transformations on the term-frequency values where terms that are extremely frequent do not increase at the same rate as in TF [5]. In [10], Global factors are designed to improve precision although this might be at the expense of a drop in recall. The rationale behind these factors is that common terms are poor discriminators.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Here, an alternative approach is adopted in this study, that is, raw TF is square root (named RTF), namely replacing raw frequency tf ( t j , d k ) with italictf()tj,dk. In general, term weighting schemes using square root function–based TF factor is superior to the logarithmic function–based TF factor . Besides, the inverse exponential frequency ( eitalicdf()tjfalse/N) in Equation can be regarded as an adjustment coefficient to reduce TF appropriately.…”
Section: Improved Weighting Scheme and Its Various Variantsmentioning
confidence: 99%
“…In general, term weighting schemes using square root function-based TF factor is superior to the logarithmic function-based TF factor. 49 Besides, the inverse exponential frequency (5) can be regarded as an adjustment coefficient to reduce TF appropriately. Meanwhile, note that the formula df(t j )/N is always less than 1 because the numerator df(t j ) is always less than or equal to denominator N. For this reason, we compute the square root of IEF (named RIEF) to further reduce the value of e −df(t j )∕N , namely replacing e −df(t j )∕N with e − √ df(t j )∕N .…”
Section: Modified Variations Of Tf-ief Schemementioning
confidence: 99%
“…Feature selection is a nontrivial preprocessing technique that alleviates the problem of high dimensionality. It reduces the number of features by counting the overall frequencies, 13 or by considering classes overlapping, 14 or by using denoising autoencoders, 15 and so forth. A more accurate technique is presented in this article to reduce the original text features into a most relevant subset of significant terms.…”
Section: Introductionmentioning
confidence: 99%