2016
DOI: 10.1007/978-3-319-25226-1_34
|View full text |Cite
|
Sign up to set email alerts
|

An Ensemble of Optimal Trees for Class Membership Probability Estimation

Abstract: General rightsThis document is made available in accordance with publisher policies. Please cite only the published version using the reference above. Abstract. Machine learning methods can be used for estimating the class membership probability of an observation. We propose an ensemble of optimal trees in terms of their predictive performance. This ensemble is formed by selecting the best trees from a large initial set of trees grown by random forest. A proportion of trees is selected on the basis of their in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
5
2

Relationship

3
4

Authors

Journals

citations
Cited by 20 publications
(36 citation statements)
references
References 12 publications
0
36
0
Order By: Relevance
“…This method requires n > p . Finally, we compare with two related ensemble methods: optimal tree ensembles (OTEs) (Khan et al ., ) and ensemble of subset of k ‐nearest‐neighbour classifiers, ES k nn (Gul et al ., ).…”
Section: Empirical Analysismentioning
confidence: 99%
“…This method requires n > p . Finally, we compare with two related ensemble methods: optimal tree ensembles (OTEs) (Khan et al ., ) and ensemble of subset of k ‐nearest‐neighbour classifiers, ES k nn (Gul et al ., ).…”
Section: Empirical Analysismentioning
confidence: 99%
“…Numbers after the abbreviations mean the training set of size of n (a subsample of the data), and then, the remaining data formed the test set. ES k NN, ensemble of subset of k NN classifiers; GP, Gaussian process; k NN, k ‐nearest neighbors; LDA, linear discriminant analysis; NSC, nearest shrunken centroids; OTE, optimal tree ensemble; PenLDA, penalized LDA; QDA, quadratic discriminant analysis; RF, random forest; RP, random projection; SRD, sum of ranking differences; SVM, support vector machine; the number after the abbreviations means “sufficient dimension reduction (SDR5) assumption”…”
Section: Resultsmentioning
confidence: 99%
“…Each random projection can be considered as a perturbation of the original data, and it is thought that the “stable” effects that are sought by statisticians are found. They focused their research on the classification performance of RPEC, and misclassification rates were selected for comparison of the following classifiers: base ones—linear discriminant analysis (LDA), quadratic discriminant analysis (QDA), k ‐nearest neighbor ( k NN), random forest (RF), support vector machines (SVMs), Gaussian process (GP) classifiers, penalized LDA (PenLDA), nearest shrunken centroids (NSCs), optimal tree ensembles (OTEs), and ensemble of subset of k NN classifiers (ES k NN). Linear and radial basis function has also been used with GP and SVM.…”
Section: Methodsmentioning
confidence: 99%
“…The problem, how diversity in an ensemble or distances between classifiers can best be measured, is discussed extensively [15 -19]. As the connection between several proposed diversity measures and the perfor mance of the ensemble is not as straightforward as might be hoped [16,19] and not always higher diversity leads to an improved ensemble [20], the classification performance of single classifiers also was considered for ensemble pruning [21][22][23].…”
Section: Discussionmentioning
confidence: 99%