2021
DOI: 10.5815/ijitcs.2021.06.05
|View full text |Cite
|
Sign up to set email alerts
|

Performance of Machine Learning Algorithms with Different K Values in K-fold CrossValidation

Abstract: The numerical value of k in a k-fold cross-validation training technique of machine learning predictive models is an essential element that impacts the model’s performance. A right choice of k results in better accuracy, while a poorly chosen value for k might affect the model’s performance. In literature, the most commonly used values of k are five (5) or ten (10), as these two values are believed to give test error rate estimates that suffer neither from extremely high bias nor very high variance. However, t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 52 publications
(34 citation statements)
references
References 0 publications
0
22
0
1
Order By: Relevance
“…We also used a 10 k‐fold cross‐validation method to further reduce instances of overfitting by testing the predictive accuracy of the training dataset (Leathwick et al, 2006). The 10 k‐fold cross‐validation works by splitting the training data into 10 equal parts and then leaving one part out each time the model is run to test the predictive performance, which can indicate selection bias or overfitting in the model (Nti et al, 2021). The BRT models were run using R statistical software (Team, 2014) and the “gbm” package (Ridgeway, 2007).…”
Section: Methodsmentioning
confidence: 99%
“…We also used a 10 k‐fold cross‐validation method to further reduce instances of overfitting by testing the predictive accuracy of the training dataset (Leathwick et al, 2006). The 10 k‐fold cross‐validation works by splitting the training data into 10 equal parts and then leaving one part out each time the model is run to test the predictive performance, which can indicate selection bias or overfitting in the model (Nti et al, 2021). The BRT models were run using R statistical software (Team, 2014) and the “gbm” package (Ridgeway, 2007).…”
Section: Methodsmentioning
confidence: 99%
“…When performing cross validation, It is typically a standard that the chosen number of folds is equal to 10 ( k = 10). 45 …”
Section: Proposed Methodology and Implementationmentioning
confidence: 99%
“…The hypothesis resulting from such operation would be the final answer. When performing cross validation, It is typically a standard that the chosen number of folds is equal to 10 ( k = 10) …”
Section: Proposed Methodology and Implementationmentioning
confidence: 99%
See 1 more Smart Citation
“…Sebagian besar penggungaan fold=10 dapat memberikan hasil yang baik tetapi terkadang di beberapa kasus fold=5 sudah cukup memadai [32]. Pada penelitian yang dilakukan Nti [33], nilai fold=7 dapat memberikan peningkatan dalam akurasi validasi. Selain itu penelitian yang dilakukan Tempola [34] dibagi kedalam 3 fold disetiap metode klasifikasi diperoleh perbandingan akurasi sistem rata-rata tertinggi.…”
Section: K-fold Cross Validationunclassified