2010
DOI: 10.1016/j.patcog.2010.02.025
|View full text |Cite
|
Sign up to set email alerts
|

A new node splitting measure for decision tree construction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
32
0
3

Year Published

2011
2011
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 71 publications
(39 citation statements)
references
References 11 publications
0
32
0
3
Order By: Relevance
“…Each decision tree in the random forest is constructed as a binary tree, its training set is produced by the bootstrapping technique and one sample can be repeated at most two times in each training set. To value the variable's splitting effect, we adopt distinct class based splitting measure [31], which consists of two parts: one is related to the number of sample classes in each child partition and the other is related to each class's proportion in the partition. To find meaningful metabolites which can discriminate liver diseases from the control, discriminate among HCC, CHB and CIR, further discriminate between each two different liver diseases, RF-RFE was conducted on five cases, labeled with RF-RFEi (i01, 2, 3, 4, and 5) in sequence.…”
Section: Discussionmentioning
confidence: 99%
“…Each decision tree in the random forest is constructed as a binary tree, its training set is produced by the bootstrapping technique and one sample can be repeated at most two times in each training set. To value the variable's splitting effect, we adopt distinct class based splitting measure [31], which consists of two parts: one is related to the number of sample classes in each child partition and the other is related to each class's proportion in the partition. To find meaningful metabolites which can discriminate liver diseases from the control, discriminate among HCC, CHB and CIR, further discriminate between each two different liver diseases, RF-RFE was conducted on five cases, labeled with RF-RFEi (i01, 2, 3, 4, and 5) in sequence.…”
Section: Discussionmentioning
confidence: 99%
“…The first gene, criterion, is an integer that indexes one of the 15 splitting criteria we have implemented: information gain [29], Gini index [5], global mutual information [19], G statistics [26], Mantaras criterion [13], hypergeometric distribution [25], Chandra-Varghese criterion [10], DCSM [9], χ 2 [27], mean posterior improvement [33], normalized gain [21], orthogonal criterion [17], twoing [5], CAIR [11] and gain ratio [32].…”
Section: Split Genesmentioning
confidence: 99%
“…Examples are the Gini index [5], the twoing criterion [5], the orthogonality criterion [17], among others. We have also included lesser-known criteria such as CAIR [11] and mean posterior improvement [33], as well as the more recent Chandra-Varghese [10] and DCSCM [9], to enhance the diversity of options for generating splits in a decision tree.…”
Section: Split Genesmentioning
confidence: 99%
“…The best attributes can be calculated through Gauss's theorem according to the available data samples. Mahmood, Imran and Satuluri [5] think that the introduction of Gini index (RGI), a new heuristic function for dimensionality reduction, could reduce the dimension of data and improve the accuracy of data classification. Chen and Xia [6] use the decision tree algorithm under the categorical variables, and introduce an optimized splitting function that constructs binary trees using domain variables, through which it can extract and select the prediction criterion.…”
Section: Introductionmentioning
confidence: 99%