2017 International Conference on Frontiers of Information Technology (FIT) 2017
DOI: 10.1109/fit.2017.00022
|View full text |Cite
|
Sign up to set email alerts
|

Resolving Class Imbalance and Feature Selection in Customer Churn Dataset

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 16 publications
0
3
0
Order By: Relevance
“…These techniques have been executed and evaluated on random forest classification model. The results illustrate that random oversampling is a better to balanced dataset [9]. Also resolved the unbalanced class problem to predict atrial fibrillation in the obese patient by utilize the SMOTE.…”
Section: Related Workmentioning
confidence: 90%
“…These techniques have been executed and evaluated on random forest classification model. The results illustrate that random oversampling is a better to balanced dataset [9]. Also resolved the unbalanced class problem to predict atrial fibrillation in the obese patient by utilize the SMOTE.…”
Section: Related Workmentioning
confidence: 90%
“…With the ensemble methods showing their strength in improving the classification performances, researchers also investigated the combinations of ensemble learning methods -especially bagging and random forests-with resampling methods to tackle the imbalance churn problem. The findings in [28], [29] show that random oversampling combined with random forests yields better results than resampling with random under-sampling and SMOTE. On the other hand, Zhu et al [30] state that bagging and random forests achieve the most promising results with respect to the profit-based measure when there is no resampling involved.…”
Section: Previous Work On Imbalanced Churn Problemmentioning
confidence: 98%
“…Support vector machines (SVM) is a powerful nonparametric supervised learning method based on structural risk minimization [28] and is proven to show good performance especially for binary classification tasks. SVM searches for the optimal separating hyperplane, which is usually a nonlinear decision function 𝑓(𝑥) = (𝑤 𝑇 𝜙(𝑥) + 𝑏), where 𝜙(𝑥) is a nonlinear transform function.…”
Section: Support Vector Machines As a Benchmarkmentioning
confidence: 99%
“…On the other hand, there is an abundance of classification models proposed for churn prediction, including: Support Vector Machines, Naïve Bayes, Decision Trees and Neural Networks [9]; Support Vector Data Description (SVDD) with random under-sampling and SMOTE oversampling [10]; combinations of random under-sampling and boosting algorithm [11]; random forest combined with random oversampling [12]; Multilayer Perceptron (MLP) neural network [13]; Reverse Nearest Neighborhood and One Class support vector machine (OCSVM) [14]; hybrid combination of well known oversampling technique SMOTE with under-sampling technique [15]; ensemble learning [16] and transfer learning methods [17]. Both complaints and churn are relatively rare events, and building statistical patterns to predict them is extremely difficult due to the imbalance of the data sets: one class (the complaints/churn) is much smaller than the other classes.…”
Section: State Of the Artmentioning
confidence: 99%