Proceedings. International Conference on Machine Learning and Cybernetics
DOI: 10.1109/icmlc.2002.1167443
|View full text |Cite
|
Sign up to set email alerts
|

Relative term-frequency based feature selection for text categorization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

9
530
0
16

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 373 publications
(555 citation statements)
references
References 9 publications
9
530
0
16
Order By: Relevance
“…Yang et al [5] has investigated several feature selections for text classification. They found that information gain and χ 2 statistic is most effective on English text dataset among five feature selection methods.…”
Section: Modified χ 2 Statisticmentioning
confidence: 99%
See 1 more Smart Citation
“…Yang et al [5] has investigated several feature selections for text classification. They found that information gain and χ 2 statistic is most effective on English text dataset among five feature selection methods.…”
Section: Modified χ 2 Statisticmentioning
confidence: 99%
“…A sophisticated methodology to reduce feature dimensionality is feature selection [5], such as χ 2 statistic, mutual information and information gain. In [6], they show introduce our term projection method in section 3.…”
Section: Introductionmentioning
confidence: 99%
“…When classifying texts, words included in them are used as classification features [21]. Undoubtedly, Markovian models are now regarded as one of the most significant state-of-the-art approaches for sequence learning.…”
Section: Hidden Markov Modelmentioning
confidence: 99%
“…In addition, most studies do not use SVM as the classification algorithm. For (1) (2) instance, Yang [12] and Pedersen [21] use kNN, and Mladenic and Grobelnic [22] use Naive Bayes [31] in their studies on keyword selection metrics. Later studies reveal that SVM performs consistently better than these classification algorithms.…”
Section: Hidden Markov Modelmentioning
confidence: 99%
“…Therefore, following Yang and Pedersen (1997), for each question we calculate the information gain of each feature of these types on the training set. We then remove those features having the lowest information gain as well as those features occurring less than ten times in the dataset.…”
Section: Feature Selectionmentioning
confidence: 99%