2017
DOI: 10.3390/a10040124
|View full text |Cite
|
Sign up to set email alerts
|

Improvement of ID3 Algorithm Based on Simplified Information Entropy and Coordination Degree

Abstract: Abstract:The decision tree algorithm is a core technology in data classification mining, and ID3 (Iterative Dichotomiser 3) algorithm is a famous one, which has achieved good results in the field of classification mining. Nevertheless, there exist some disadvantages of ID3 such as attributes biasing multi-values, high complexity, large scales, etc. In this paper, an improved ID3 algorithm is proposed that combines the simplified information entropy based on different weights with coordination degree in rough s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 23 publications
(10 citation statements)
references
References 32 publications
0
9
0
1
Order By: Relevance
“…For a dataset with one class label, will be 1 and ( ) is 0. Hence the Entropy of homogenous data set is zero [8]. If the entropy is higher the uncertainty/impurity/mixing is higher [9].…”
Section: A Entropymentioning
confidence: 99%
See 1 more Smart Citation
“…For a dataset with one class label, will be 1 and ( ) is 0. Hence the Entropy of homogenous data set is zero [8]. If the entropy is higher the uncertainty/impurity/mixing is higher [9].…”
Section: A Entropymentioning
confidence: 99%
“…The feature with highest information gain is the best feature to be selected for split. Assuming that there are V different values for a feature f, |L v | represents the subset of L with f=v, Information gain after splitting L on a feature f is measured as [8].…”
Section: B Information Gainmentioning
confidence: 99%
“…Pemodelan klasifikasi adalah proses ekstraksi data kesesuaian lahan yang telah ada [18], yang dalam algoritma SDT menggunakan entropi untuk menumbuhkan pohon keputusan. Berdasarkan hal tersebut, suatu variabel dalam suatu dataset yang memiliki sifat heterogenitas yang tinggi menyebakan nilai entropinya tinggi, dan sebaliknya jika data dalam suatu variabel bersifat homogen maka nilai entropinya akan rendah atau bahkan bernilai 0 [28]. Nilai entropi yang tinggi pada suatu variabel menyebabkan variabel tersebut memiliki peluang yang kecil untuk dijadikan sebagai simpul akar/ internal, seperti halnya variabel drainase dan kapasitas tukar kation.…”
Section: Evaluasi Hasil Klasifikasiunclassified
“…A hyperparameter is an internal parameter of a classifier method, such as the box constraint of DT or a support vector machine, or maybe the learning rate of a robust classification ensemble. The goal of these parameters can highly affect the performance of a classifier; BO uses a fit function [9]. For Model selection, depend on Hyper-parameters, features, train a classifier like a decision tree, using a dataset, this process needs to choose the best feature set and hyperparameters by applying K fold DT.…”
Section: Dt Algorithms With Hyperparameter Optimizationmentioning
confidence: 99%