2020
DOI: 10.14569/ijacsa.2020.0110277
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating the Impact of GINI Index and Information Gain on Classification using Decision Tree Classifier Algorithm*

Abstract: Decision tree is a supervised machine learning algorithm suitable for solving classification and regression problems. Decision trees are recursively built by applying split conditions at each node that divides the training records into subsets with output variable of same class. The process starts from the root node of the decision tree and progresses by applying split conditions at each non-leaf node resulting into homogenous subsets. However, achieving pure homogenous subsets is not possible. Therefore, the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
86
0
9

Year Published

2020
2020
2024
2024

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 148 publications
(96 citation statements)
references
References 14 publications
1
86
0
9
Order By: Relevance
“…Another way that shows the importance of a feature is to rank the tree due to the decrease in impurity (Gini impurity) relative to all trees [56]. According to the principle of the algorithm, the most impure trees are at the beginning and the least impure trees are at the end.…”
Section: Feature Selectionmentioning
confidence: 99%
“…Another way that shows the importance of a feature is to rank the tree due to the decrease in impurity (Gini impurity) relative to all trees [56]. According to the principle of the algorithm, the most impure trees are at the beginning and the least impure trees are at the end.…”
Section: Feature Selectionmentioning
confidence: 99%
“…Gini impurity adalah ukuran seberapa sering elemen yang dipilih secara acak dari himpunan akan diberi label yang salah jika diberi label secara acak sesuai dengan distribusi label di dalam subset. Penggunaan indeks gini dan perolehan informasi sebelum maupun sesudah data set balanced tidak mempengaruhi performa model [9].…”
Section: Gini Impurityunclassified
“…There are different selection measures for the splitting criteria of the decision tree that performs the data portioning into best possible manner. Information Gain, Gain Ratio, and Gini Index are most popular examples (Tangirala, 2020). We have measured our evaluation with Information Gain and Gini Index.…”
Section: G Decision Treementioning
confidence: 99%