2020
DOI: 10.1016/j.knosys.2019.06.034
|View full text |Cite
|
Sign up to set email alerts
|

Improving interpolation-based oversampling for imbalanced data learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
2
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 56 publications
(19 citation statements)
references
References 44 publications
0
18
0
Order By: Relevance
“…Other ways of oversampling include, but are not limited to, the work of [91,92,93,94,78,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119] The validation process is what all oversampling methods have in common, which is basically the evaluation of the classifier's performance employed to classify the oversampled datasets using one or more accuracy measures such as Accuracy, Precision, Recall, F-measure, G-mean, Specificity, Kappa, Matthews correlation coefficient (MCC), Area under the ROC Curve (AUC), True positive rate, False negative (FN), False positive (FP), True positive (TP), True negative (TN), and ROC curve. Table 1 lists 72 oversampling methods, including their known names, references, the number of datasets utilized, the number of classes in these datasets, the classifiers employed, and the performance metrics used to validate the classification results after oversampling.…”
Section: Literature Review Of Oversampling Methodsmentioning
confidence: 99%
“…Other ways of oversampling include, but are not limited to, the work of [91,92,93,94,78,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119] The validation process is what all oversampling methods have in common, which is basically the evaluation of the classifier's performance employed to classify the oversampled datasets using one or more accuracy measures such as Accuracy, Precision, Recall, F-measure, G-mean, Specificity, Kappa, Matthews correlation coefficient (MCC), Area under the ROC Curve (AUC), True positive rate, False negative (FN), False positive (FP), True positive (TP), True negative (TN), and ROC curve. Table 1 lists 72 oversampling methods, including their known names, references, the number of datasets utilized, the number of classes in these datasets, the classifiers employed, and the performance metrics used to validate the classification results after oversampling.…”
Section: Literature Review Of Oversampling Methodsmentioning
confidence: 99%
“…This means that a better model can be obtained by creating a model by expanding the experimental data than the modeling result using 18 data elements, which is the actual data. In fact, existing studies have reported that the creation of virtual data improves the R 2 and MSE values of the model [47][48][49].…”
Section: Evaluation Of the Modelmentioning
confidence: 99%
“…Many extensions and modifications to the SMOTE algorithm have been proposed, e.g., the Borderline SMOTE [28] (bSMOTE), which detects and uses the borderline observations to generate new synthetic samples. Other over-sampling concepts employ clusterization [14], genetic algorithms [29], potential functions [30], neighborhood analysis [30], interpolation [31], and identification of harder to learn observations [32].…”
Section: Related Workmentioning
confidence: 99%