2013
DOI: 10.1007/978-1-4614-6849-3
|View full text |Cite
|
Sign up to set email alerts
|

Applied Predictive Modeling

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

12
3,334
1
122

Year Published

2015
2015
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 3,643 publications
(3,469 citation statements)
references
References 0 publications
12
3,334
1
122
Order By: Relevance
“…For each study area, random forests (Liaw and Wiener, 2002; parameters mtry = default and 299 ntree = 1000) was used to calculate covariate importance, as random forests is not highly sensitive to 300 non-informative predictors (Kuhn and Johnson, 2013). Random forests identifies important covariates 301 by generating multiple classification trees (a forest) using bootstrap sampling, randomly scrambling the 302 covariates in each bootstrap sample and reclassifying the bootstrap sample.…”
Section: Utah Clhs 207 208mentioning
confidence: 99%
“…For each study area, random forests (Liaw and Wiener, 2002; parameters mtry = default and 299 ntree = 1000) was used to calculate covariate importance, as random forests is not highly sensitive to 300 non-informative predictors (Kuhn and Johnson, 2013). Random forests identifies important covariates 301 by generating multiple classification trees (a forest) using bootstrap sampling, randomly scrambling the 302 covariates in each bootstrap sample and reclassifying the bootstrap sample.…”
Section: Utah Clhs 207 208mentioning
confidence: 99%
“…The predictive power in the data may depend significantly on the way missing values are treated. While some machine learning algorithms, such as decision trees [16], have the capability to handle missing data outright, most machine learning algorithms do not. In many situations missing values are imputed using a supervised learning technique such as k-Nearest Neighbour (KNN) after suitable scaling to balance the contribution of the numeric attributes.…”
Section: Imputationmentioning
confidence: 99%
“…These imputation techniques do not have theoretical formulations but have been much implemented in practice [4] [6]. In this work, we considered different imputations such as the KNN imputation, the tree bagging imputation from the caret package [16], and the random forest imputation from the randomForest package [17]. The last method led to the best results in terms of the performance of the predictive models finally built, although it was more computationally expensive.…”
Section: Imputationmentioning
confidence: 99%
See 2 more Smart Citations