2022
DOI: 10.1109/access.2022.3169512
|View full text |Cite
|
Sign up to set email alerts
|

Stop Oversampling for Class Imbalance Learning: A Review

Abstract: For the last two decades, oversampling has been employed to overcome the challenge of learning from imbalanced datasets. Many approaches to solving this challenge have been offered in the literature. Oversampling, on the other hand, is a concern. That is, models trained on fictitious data may fail spectacularly when put to real-world problems. The fundamental difficulty with oversampling approaches is that, given a real-life population, the synthesized samples may not truly belong to the minority class. As a r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

1
24
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 53 publications
(26 citation statements)
references
References 178 publications
1
24
0
1
Order By: Relevance
“…The number of nearest neighbors is a hyper-parameter for SMOTE and is usually selected based on performance over a specific metric, e.g., model accuracy on the validation set. SMOTE, however, does not consider the underlying class distribution and is prune to over-fitting, class overlap, and noisy sample generation [14]. Although a number of potential improvements over SMOTE have been proposed, (see [14] and references therein), over time, SMOTE has become more popular among researchers and has been seen as the default upsampling method in the field of imbalanced learning.…”
Section: Appendix a Synthetic Minority Oversampling Techniquementioning
confidence: 99%
See 3 more Smart Citations
“…The number of nearest neighbors is a hyper-parameter for SMOTE and is usually selected based on performance over a specific metric, e.g., model accuracy on the validation set. SMOTE, however, does not consider the underlying class distribution and is prune to over-fitting, class overlap, and noisy sample generation [14]. Although a number of potential improvements over SMOTE have been proposed, (see [14] and references therein), over time, SMOTE has become more popular among researchers and has been seen as the default upsampling method in the field of imbalanced learning.…”
Section: Appendix a Synthetic Minority Oversampling Techniquementioning
confidence: 99%
“…SMOTE, however, does not consider the underlying class distribution and is prune to over-fitting, class overlap, and noisy sample generation [14]. Although a number of potential improvements over SMOTE have been proposed, (see [14] and references therein), over time, SMOTE has become more popular among researchers and has been seen as the default upsampling method in the field of imbalanced learning. Thus, we chose to include SMOTE as a benchmark method in our analysis.…”
Section: Appendix a Synthetic Minority Oversampling Techniquementioning
confidence: 99%
See 2 more Smart Citations
“…There are a number of AE variants in literature [14], out of which the Variational Autoencoders (VAEs) [12] are the most popular. In vanilla AE, the encoder outputs a single latent vector z directly, however, in VAEs, the encoder outputs two vectors; a mean vector µ and a variance vector σ.…”
Section: Appendix B Variational Autoencodersmentioning
confidence: 99%