2004
DOI: 10.1016/s0020-0255(03)00414-6
|View full text |Cite
|
Sign up to set email alerts
|

A hybrid decision tree/genetic algorithm method for data mining

Abstract: This paper addresses the well-known classification task of data mining, where the objective is to predict the class which an example belongs to. Discovered knowledge is expressed in the form of high-level, easy-to-interpret classification rules. In order to discover classification rules, we propose a hybrid decision tree/genetic algorithm method. The central idea of this hybrid method involves the concept of small disjuncts in data mining, as follows. In essence, a set of classification rules can be regarded a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
29
0
1

Year Published

2011
2011
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(30 citation statements)
references
References 7 publications
(11 reference statements)
0
29
0
1
Order By: Relevance
“…[79] Table 6 concerns results from [79], where 10 different methods of automatic classifier construction based on data were considered. We chose 7 out of them for comparative analysis: C45R (C45Rules technique of generating a set M a n u s c r i p t of production rules from a tree produced by C4.5 algorithm) [83], RIPPER (improvement of the efficient Incremental Reduced Error Pruning algorithm) [84], MPLCS (a Memetic Pittsburgh Learning Classifier System) [85], AntMin+ (Ant Colony-based data miner to extract classification rules inspired by the research on the behaviour of real ant colonies) [86], CORE (COevolutionary Rule Extractor) [87], DTGA (hybrid decision tree/genetic algorithm discovering rules on small disjuncts) [88], and GFS-GP (Genetic Programming algorithm used to learn fuzzy rule-based classifiers) [89]. Three remaining methods of [79] were skipped: ILGA (Incremental Learning approach to Genetic Algorithms) [90], SLAVE (Structural Learning Algorithm on Vague Environment) [91], and the ICRM (Interpretable Classification Rule Mining algorithm) method proposed by the authors of [79].…”
Section: A Comparative Analysis With Alternative Techniquesmentioning
confidence: 99%
“…[79] Table 6 concerns results from [79], where 10 different methods of automatic classifier construction based on data were considered. We chose 7 out of them for comparative analysis: C45R (C45Rules technique of generating a set M a n u s c r i p t of production rules from a tree produced by C4.5 algorithm) [83], RIPPER (improvement of the efficient Incremental Reduced Error Pruning algorithm) [84], MPLCS (a Memetic Pittsburgh Learning Classifier System) [85], AntMin+ (Ant Colony-based data miner to extract classification rules inspired by the research on the behaviour of real ant colonies) [86], CORE (COevolutionary Rule Extractor) [87], DTGA (hybrid decision tree/genetic algorithm discovering rules on small disjuncts) [88], and GFS-GP (Genetic Programming algorithm used to learn fuzzy rule-based classifiers) [89]. Three remaining methods of [79] were skipped: ILGA (Incremental Learning approach to Genetic Algorithms) [90], SLAVE (Structural Learning Algorithm on Vague Environment) [91], and the ICRM (Interpretable Classification Rule Mining algorithm) method proposed by the authors of [79].…”
Section: A Comparative Analysis With Alternative Techniquesmentioning
confidence: 99%
“…There are many studies trying to improve classifying efficiency. Chang and Liu [11] and Carvalho and Freitas [8] applied decision tree for other method such as K-mean, fuzzy C mean. Chandra and Varghese [10] proposed a binary decision tree algorithm using the Gini index as a split measure.…”
Section: Fuzzy Decision Tree (Fdt)mentioning
confidence: 99%
“…Total cost (TC) is by defining and discussing the links between R&D cost, market capital, and design quality for market share in Eq. (8). Market share can be gained by attracting customers with preferences more distant from the target market.…”
Section: The Formulation Of Product Mix-experience Problems Considerimentioning
confidence: 99%
“…The most common strategy is to evaluate every possible subtree, working from the leaves backward, for possible replacement by a terminal node. Nontrivial heuristics including genetic algorithms [114] have also been demonstrated as useful pruning strategies [115,116].…”
Section: Decision Trees and Random Forestsmentioning
confidence: 99%