2022
DOI: 10.1016/j.infsof.2021.106736
|View full text |Cite
|
Sign up to set email alerts
|

A comparison of machine learning algorithms on design smell detection using balanced and imbalanced dataset: A study of God class

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
43
2
2

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 28 publications
(47 citation statements)
references
References 20 publications
0
43
2
2
Order By: Relevance
“…Therefore, no conclusive empirical evidence from their experimental results showed that using SMOTE, RUS, and ROS could significantly positively impact machine learning-based CSD models. Alkharabsheh et al Alkharabsheh et al (2022) used the machine learning classifiers (i.e., LDA, Quadratic Discriminant Analysis (QDA), NB, Multi-Layer Perceptron (MLP), SVM, DT, GB, CatBoost, Light Gradient Boosting Machine (LGBM), XGBoost, XGBoost with Random Forest (XGBRF), AdaBoost, Bagging, RF, Extra Trees (ET), KNN, Nearest Centroid (NC), Gaussian Process (GP), Ridge, LR, Perceptron, Passive Aggressive (PA), and Stochastic Gradient Descent (SGD)) to compare whether using SMOTE would improve the detection performance on God Class detection. Their results showed that SMOTE could not improve the God Class detection performance.…”
Section: Imbalanced Learning For Csdmentioning
confidence: 99%
See 3 more Smart Citations
“…Therefore, no conclusive empirical evidence from their experimental results showed that using SMOTE, RUS, and ROS could significantly positively impact machine learning-based CSD models. Alkharabsheh et al Alkharabsheh et al (2022) used the machine learning classifiers (i.e., LDA, Quadratic Discriminant Analysis (QDA), NB, Multi-Layer Perceptron (MLP), SVM, DT, GB, CatBoost, Light Gradient Boosting Machine (LGBM), XGBoost, XGBoost with Random Forest (XGBRF), AdaBoost, Bagging, RF, Extra Trees (ET), KNN, Nearest Centroid (NC), Gaussian Process (GP), Ridge, LR, Perceptron, Passive Aggressive (PA), and Stochastic Gradient Descent (SGD)) to compare whether using SMOTE would improve the detection performance on God Class detection. Their results showed that SMOTE could not improve the God Class detection performance.…”
Section: Imbalanced Learning For Csdmentioning
confidence: 99%
“…The former produces a superset of the original code smell data sets by duplicating existing smelly instances or creating new smelly instances from existing smelly ones, while the latter produces a subset of the original code smell data sets by eliminating non-smelly instances. To maintain consistency with previous studies Pecorelli et al (2020); Alkharabsheh et al (2022) and common practices, we set the default smelly ratio to 0.5, resulting in an equal number of smelly and non-smelly instances in the balanced data sets.…”
Section: Data Resamplingmentioning
confidence: 99%
See 2 more Smart Citations
“…The main role of software metrics is to estimate and measure some characteristics of systems such as size, complexity, inheritance, encapsulation, etc. [14,15]. Selected metrics are a large set of object-oriented metrics that are considered as independent variables as shown in Table 1.…”
Section: Introductionmentioning
confidence: 99%