Instance Selection and Construction for Data Mining 2001
DOI: 10.1007/978-1-4757-3359-4_6
|View full text |Cite
|
Sign up to set email alerts
|

Genetic-Algorithm-Based Instance and Feature Selection

Abstract: This chapter discusses a genetic-algorithm-based approach for selecting a small number of instances from a given data set in a pattern classification problem. Our genetic algorithm also selects a small number of features. The selected instances and features are used as a reference set in a nearest neighbor classifier. Our goal is to improve the classification ability of our nearest neighbor classifier by searching for an appropriate reference set. We first describe the implementation of our genetic algorithm f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2003
2003
2013
2013

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 19 publications
(13 citation statements)
references
References 9 publications
(12 reference statements)
0
13
0
Order By: Relevance
“…We test the suitability of the approach on di↵erent datasets and compare the performance achieved to that of existing CBM algorithms from the literature. Previous works are mainly focused on reducing either the number of redundant cases or noisy cases [1,7,8,[20][21][22], or aimed at selecting attributes [10,12] or to both enhance the accuracy and reduce the size of the case-base [11]. However, the fitness function proposed in this work measures the redundancy of the case-base, the number of noisy cases and the error rate of the system.…”
Section: Discussionmentioning
confidence: 99%
“…We test the suitability of the approach on di↵erent datasets and compare the performance achieved to that of existing CBM algorithms from the literature. Previous works are mainly focused on reducing either the number of redundant cases or noisy cases [1,7,8,[20][21][22], or aimed at selecting attributes [10,12] or to both enhance the accuracy and reduce the size of the case-base [11]. However, the fitness function proposed in this work measures the redundancy of the case-base, the number of noisy cases and the error rate of the system.…”
Section: Discussionmentioning
confidence: 99%
“…An initial population of chromosomes are selected either randomly or with some heuristics, and the selection-crossover-mutation operations are applied to the chromosomes to obtain better solutions. Ishibuchi [13] attempted to apply the GA to both feature and instance selection simultaneously and did obtain some good results, but the study is limited to the problem domain where the nearest neighbor method is applied for classification analysis on numeric data. In short, it appears that existing studies on applying the GA for data reduction tend to be task-specific (classification in particular), and whether and how the GA techniques can be used for general purpose data reduction remains to be investigated.…”
Section: Adaptive Sampling Proceduresmentioning
confidence: 99%
“…The application of the GAs for feature selection has been explored in [5,13,23], among others. Since feature selection is often associated with classification problems, the fitness functions specified in these studies are more or less a function of classification error rate.…”
Section: Adaptive Sampling Proceduresmentioning
confidence: 99%
See 1 more Smart Citation
“…Finally, a related area of research is that of the use of GAs for feature and instance selection in, for example, data mining. A number of references to this literature can be found in [10].…”
Section: Introductionmentioning
confidence: 99%