2021
DOI: 10.1016/j.eswa.2021.114582
|View full text |Cite
|
Sign up to set email alerts
|

Conditional Wasserstein GAN-based oversampling of tabular data for imbalanced learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
59
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 146 publications
(75 citation statements)
references
References 19 publications
0
59
0
Order By: Relevance
“…The imbalanced classification performance of our approach is compared against four commonly used resampling method, i.e., CCR [5], k SMOTE [3], GAN [17] and CUSBoost [18]. The former two are oversampling, while the last one is undersampling.…”
Section: Experimental Set-upsmentioning
confidence: 99%
“…The imbalanced classification performance of our approach is compared against four commonly used resampling method, i.e., CCR [5], k SMOTE [3], GAN [17] and CUSBoost [18]. The former two are oversampling, while the last one is undersampling.…”
Section: Experimental Set-upsmentioning
confidence: 99%
“…To this end, we propose a technique to generate realistic botnet data using Generative Adversarial Networks (GANs) to improve the classifiers' decision making to detect potential evasion samples. GANs have proved to be highly effective in some recent research works [6]- [10]. A GAN is a combination of two different AI models competitively learning to generate realistic samples.…”
Section: Introductionmentioning
confidence: 99%
“…In such scenarios, the machine learning classifiers over-fit the majority class and fall short in generalizing the test set [12]. The motivation for using GANs for data oversampling is their effectiveness in mimicking complex probably distributions [10]. To address the low-data regime problem, synthetic oversampling techniques like SMOTE [13] are employed, but these techniques depend on algorithms like nearest neighbours and linear interpolation which make them unsuitable for highdimensional and complex probability distributions of data [10].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The basic idea of GAN is to use a generator to generate samples that people need from random data points that meet a specific distribution (for example, Gaussian distribution). Some scholars use the ability of GAN to learn images and apply it to the field of image anomaly detection, such as AnoGAN [36], BiGAN [37] and GANomaly [38] and some GAN-based imbalanced data intrusion detection models [61]- [63]. These GAN-based network architectures have shown high performance.…”
Section: Introductionmentioning
confidence: 99%