2020
DOI: 10.1109/access.2020.3028469
|View full text |Cite
|
Sign up to set email alerts
|

A Fast Non-Redundant Feature Selection Technique for Text Data

Abstract: Feature selection is critical in reducing the size of data and improving classifier accuracy by selecting an optimum subset of the overall features. Traditionally, each feature is given a score against a particular category (such as using Mutual Information) and the task of feature selection comes down to choosing the top k ranked features with the best average score across all categories. However, this approach has two major drawbacks. Firstly, the maximum or average score of a feature with a class might not … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 21 publications
(7 citation statements)
references
References 56 publications
(76 reference statements)
0
7
0
Order By: Relevance
“…The within-class scatter matrix Wid class and between-class scatter matrix Bet class is being used to overcome the problems associated with the ideal discrimination projection matrix. Furthermore, Equation (11) depicts the mathematical equation used to calculate the projection matrix. In addition, the formulas for evaluating Bet class and Wid class are represented by the mathematical equations Equation (12) and Equation (13), respectively.…”
Section: Sigmoid Kernelmentioning
confidence: 99%
See 3 more Smart Citations
“…The within-class scatter matrix Wid class and between-class scatter matrix Bet class is being used to overcome the problems associated with the ideal discrimination projection matrix. Furthermore, Equation (11) depicts the mathematical equation used to calculate the projection matrix. In addition, the formulas for evaluating Bet class and Wid class are represented by the mathematical equations Equation (12) and Equation (13), respectively.…”
Section: Sigmoid Kernelmentioning
confidence: 99%
“…In addition, the formulas for evaluating Bet class and Wid class are represented by the mathematical equations Equation (12) and Equation (13), respectively. The eigenvectors of s are shown in Equation (11). The projection matrix is denoted by the symbol S and the eigenvectors of S are shown in Equation ( 14), in which G T = Bet class + Wid class ,, F J is the feature vector of the data, š›¼ N and N is the data vector and samples in the data class J.…”
Section: Sigmoid Kernelmentioning
confidence: 99%
See 2 more Smart Citations
“…This is particularly true for sparse data. Feature selection is a commonly used approach to select only relevant (discriminatory) features [8,9] which help in increasing the effectiveness of an algorithm as well as reducing its complexity. This is essential since all features are not equally important and irrelevant features negatively impact the clusters.…”
Section: Introductionmentioning
confidence: 99%