2021
DOI: 10.1016/j.ins.2021.01.020
|View full text |Cite
|
Sign up to set email alerts
|

TAGA: Tabu Asexual Genetic Algorithm embedded in a filter/filter feature selection approach for high-dimensional data

Abstract: Feature selection is the process of selecting an optimal subset of features required for maintaining or improving the performance of data mining models. Recently, hybrid filter/wrapper feature selection methods have shown promising results for high-dimensional data. However, filter/wrapper methods lack of generalisation power, which enables the selected features to be trainable over different classifiers without having to repeat the feature selection process. To address the generalisation power problem, this p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 34 publications
(12 citation statements)
references
References 35 publications
0
12
0
Order By: Relevance
“…Aside from identifying features relevant to the classification problem, feature selection reduces the dimensionality of the dataset, simplifying the problem, thereby improving model stability and generalisability. Feature selection was performed using a filter-based tabu asexual genetic algorithm (TAGA) 28 and the knowledge of clinical experts.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Aside from identifying features relevant to the classification problem, feature selection reduces the dimensionality of the dataset, simplifying the problem, thereby improving model stability and generalisability. Feature selection was performed using a filter-based tabu asexual genetic algorithm (TAGA) 28 and the knowledge of clinical experts.…”
Section: Resultsmentioning
confidence: 99%
“…A Tabu Asexual Genetic Algorithm (TAGA) 28 was used to generate 9 feature sets (Supplementary Table S3 ). TAGA takes an m × n matrix, where m is the number of samples and n is the number of features and calculates a Fisher’s score.…”
Section: Methodsmentioning
confidence: 99%
“…Tables 13 and 14 describe the comparison of experimental results between HFIA and the feature selection method mentioned in [ 56 ].…”
Section: Experiments and Discussionmentioning
confidence: 99%
“…Firstly, the collected annual report text data are preprocessed, and then unigrams, bigrams, and trigrams are extracted as text features by using word bag model and word frequency reverse document frequency (TF-IDF) weighting method. Because text features naturally face high-dimensional problems, high-dimensional text features may contain some redundant and irrelevant features [12]. erefore, the information gain method is further used to filter the extracted initial text features, and the important features are retained to ensure the quality of the features.…”
Section: Feature Extraction Of Financial Risk Predictionmentioning
confidence: 99%