2021
DOI: 10.1007/s10462-021-09970-6
|View full text |Cite
|
Sign up to set email alerts
|

Feature selection methods for text classification: a systematic literature review

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 73 publications
(38 citation statements)
references
References 189 publications
0
30
0
Order By: Relevance
“…The reason for using a CC is that it provides accurate results with reasonably fast execution. The same reasoning applies to the use of NBs as base classifiers and it is also a popular base classifier that is frequently used in the literature (Pintas et al, 2021). Furthermore, our concept drift adapting strategy resets the classifier if drift is detected.…”
Section: Experimental Setup and Evaluation Metricsmentioning
confidence: 91%
“…The reason for using a CC is that it provides accurate results with reasonably fast execution. The same reasoning applies to the use of NBs as base classifiers and it is also a popular base classifier that is frequently used in the literature (Pintas et al, 2021). Furthermore, our concept drift adapting strategy resets the classifier if drift is detected.…”
Section: Experimental Setup and Evaluation Metricsmentioning
confidence: 91%
“…In this paper, the information gain (IG) as the filter method and the genetic algorithm (GA) as the wrapper method are used for feature selection. Particularly, these two methods have been used in many research problems, including text classification [20], gene expression microarray analysis [21], intrusion detection [22], financial distress prediction [23], software defect prediction [24], etc.…”
Section: The Feature Selection and Over-sampling Methodsmentioning
confidence: 99%
“…Feature selection represents an important tool to balance the number of selected attributes to avoid overfitting the model (with too few attributes) and expensive computational time (with too many attributes). There are many methods for feature selection such as: wrapper methods [11], filter methods and unsupervised methods. Wrapper and filter methods are considered supervised approaches as they utilize the output to produce the best set of features.…”
Section: Related Workmentioning
confidence: 99%