2015
DOI: 10.12988/ams.2015.58562
|View full text |Cite
|
Sign up to set email alerts
|

SMOTE bagging algorithm for imbalanced dataset in logistic regression analysis (case: credit of bank X)

Abstract: Logistic regression analysis is one of classification methods which is both most popular and common used. This classifier works well when the class distribution in response variable is balanced. In many real cases, the imbalanced class dataset frequently was found. This problem can affect of being difficult at obtaining a good predictive model for minority class dataset. The prediction accuracy generated will be good for majority class but not for minority class. SMOTEBagging is a combination of SMOTE and Bagg… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 31 publications
(10 citation statements)
references
References 10 publications
0
8
0
Order By: Relevance
“…A thing to take note when using supervised method for training is imbalanced data: The predictive models developed using conventional machine learning algorithms could be biased and inaccurate because the number of observations in one class of the dataset is significantly lower than the other. To handle imbalanced data, several methods can be used, including resampling, boosting, bagging [17][18][19][20].…”
Section: Supervised Modelmentioning
confidence: 99%
“…A thing to take note when using supervised method for training is imbalanced data: The predictive models developed using conventional machine learning algorithms could be biased and inaccurate because the number of observations in one class of the dataset is significantly lower than the other. To handle imbalanced data, several methods can be used, including resampling, boosting, bagging [17][18][19][20].…”
Section: Supervised Modelmentioning
confidence: 99%
“…In 2014, a method based on cost-sensitive decision tree with feature space partitioning was introduced (Krawczyk et al, 2014), and the results were computed on different benchmark datasets with varying imbalance ratios (IR). The analysis of SMOTEBagging with logistic regression using credit scoring data revealed its higher degree of accuracy compared to a simple logistic algorithm (Hanifah et al, 2015). A new ensemble classification method using random undersampling and ROSE sampling under a boosting scheme RHSBoost was proposed to address the imbalance classification problem (Gong & Kim, 2017).…”
Section: Related Workmentioning
confidence: 99%
“…To overcome such limitations, several algorithms have been proposed as extensions of SMOTE. Some are focusing on improving the generation of synthetic data by combining SMOTE with other oversampling techniques, including the combination of SMOTE with Tomek-links (Elhassan et al 2016), particle swarm optimization (Gao et al 2011;Wang et al 2014), rough set theory (Ramentol et al 2012), kernel based approaches (Mathew et al 2015), Boosting (Chawla et al 2003), and Bagging (Hanifah et al 2015). Other approaches choose subsets of the minority class data to generate SMOTE samples or cleverly limit the number of synthetic data generated (Santoso et al 2017).…”
Section: Introductionmentioning
confidence: 99%