2020
DOI: 10.48550/arxiv.2008.09202
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Conditional Wasserstein GAN-based Oversampling of Tabular Data for Imbalanced Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(16 citation statements)
references
References 0 publications
0
13
1
Order By: Relevance
“…Similar to (Fiore et al 2019), the author synthesises the underrepresented class hence making known the class label. Other work has shown the efficacy of generative networks above traditional methods (Liu et al 2019;Ngwenduna and Mbuvha 2021;Engelmann and Lessmann 2020). However, we do not find any studies reporting the similarity of the synthesisers to the original dataset.…”
Section: Introductioncontrasting
confidence: 58%
“…Similar to (Fiore et al 2019), the author synthesises the underrepresented class hence making known the class label. Other work has shown the efficacy of generative networks above traditional methods (Liu et al 2019;Ngwenduna and Mbuvha 2021;Engelmann and Lessmann 2020). However, we do not find any studies reporting the similarity of the synthesisers to the original dataset.…”
Section: Introductioncontrasting
confidence: 58%
“…However, undersampling might result in the loss of diversity. For oversampling, methods like SMOTE use nearest neighbours and linear interpolation, which can be unsuitable for high-dimensional and complex probability distributions [8], [21]. Recent research works proposed algorithms for data oversampling.…”
Section: B Data Oversampling and Gansmentioning
confidence: 99%
“…Using generative adversarial networks (GANs) as synthetic oversamplers has been a voguish research endeavour for low data regimes [3], [7]. Various researchers have demonstrated that GANs are more effective as compared to other synthetic oversamplers like SMOTE [2], [6], [8], [9]. It is found in many studies that due to the adversarial factor, GANs can better estimate the target probability distribution [2], [8], [10].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…GAN's framework corresponds to a minimax two-player game, it simultaneously trains two models: a generative model G that captures the data distribution, and a discriminative model D that estimates the probability that a sample came from the training data rather than G. GAN is becoming more and more popular in the field of content generation [45,46]. In the field of credit scoring, GAN has been used to solve the sample imbalance problem [47].…”
Section: Adversarial Validationmentioning
confidence: 99%