2022
DOI: 10.33395/sinkron.v7i4.11792
|View full text |Cite
|
Sign up to set email alerts
|

Stratified K-fold cross validation optimization on machine learning for prediction

Abstract: Cervical is the second most common malignant tumor in women, with 341,000 deaths worldwide in 2020, almost 80% of which occur in developing countries. One of the causes is infection with Human papillomavirus (HPV) types 16 and 18. The increasing incidence of cervical cancer in Indonesia makes this disease must be treated seriously because it is one of the main causes of death. In addition to the virus, external factors can be one of the causes. The high mortality rate in patients is caused by the patient's awa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 15 publications
0
3
0
1
Order By: Relevance
“…Five-fold cross-validation was used, dividing the dataset into five subdivisions and taking four subdivisions each time as the training set and the remaining one sub-division as the test set. Stratified KFold (Widodo et al ., 2022 ) was used to perform 5-fold cross-validation in this study. To avoid feature leakage, feature selection was only applied to the training dataset.…”
Section: Methodsmentioning
confidence: 99%
“…Five-fold cross-validation was used, dividing the dataset into five subdivisions and taking four subdivisions each time as the training set and the remaining one sub-division as the test set. Stratified KFold (Widodo et al ., 2022 ) was used to perform 5-fold cross-validation in this study. To avoid feature leakage, feature selection was only applied to the training dataset.…”
Section: Methodsmentioning
confidence: 99%
“…Model performance measurement is carried out using a confusion matrix and 10-fold cross validation. Confusion Matrix helps to understand how well a classification model predicts correctly, while K-Fold Cross-Validation helps measure the overall performance of the model by avoiding bias caused by random separation of test and training data [22], [23].…”
Section: Methodsmentioning
confidence: 99%
“…Stratified K-Fold Cross-Validation (SKCV) is an extension of KCV, where class distribution in the original data is taken into consideration when sampling [18]. Accordingly, SKCV is preferred over KCV in the case of unbalanced class distributions [19]. In our experiments, we used SKCV, specifically 10-fold cross validation, to split the data into training and testing, while computing the average accuracy of the different folds.…”
Section: ) K-fold Cross-validation (Kcvmentioning
confidence: 99%