2021
DOI: 10.1007/978-981-16-6285-0_2
|View full text |Cite
|
Sign up to set email alerts
|

Effective Rate of Minority Class Over-Sampling for Maximizing the Imbalanced Dataset Model Performance

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 13 publications
0
2
0
Order By: Relevance
“…In the literature [11], the authors use an oversampling approach to replicate randomly selected samples from a small number, which reduces both inter-category and intra-category imbalances, but such an approach can lead to the occurrence of overfitting. In the literature [12], the authors use the undersampling method to remove randomly from the majority of classes until all classes have the same number of samples, but the significant disadvantage of this is that it discards a portion of the available data. Unlike the data level, the algorithm level is to modify the training algorithm or the network structure.…”
Section: Class Imbalance Machine Learningmentioning
confidence: 99%
“…In the literature [11], the authors use an oversampling approach to replicate randomly selected samples from a small number, which reduces both inter-category and intra-category imbalances, but such an approach can lead to the occurrence of overfitting. In the literature [12], the authors use the undersampling method to remove randomly from the majority of classes until all classes have the same number of samples, but the significant disadvantage of this is that it discards a portion of the available data. Unlike the data level, the algorithm level is to modify the training algorithm or the network structure.…”
Section: Class Imbalance Machine Learningmentioning
confidence: 99%
“…7, SVM model in the F1 score, accuracy, and precision metrics have provided the highest values, and the RF model in AUC and sensitivity has given the best amounts. We prefer the F1 score over other metrics because it is effective for imbalanced datasets [91], and our test data consists of 76% failed and 24% Not-failed conditions, which is an imbalanced dataset. After F1 score, AUC with the highest value represented the best predictive model [92].…”
Section: Evaluating the Best-trained Modelsmentioning
confidence: 99%