2017
DOI: 10.1587/transinf.2016edp7204
|View full text |Cite
|
Sign up to set email alerts
|

The Performance Stability of Defect Prediction Models with Class Imbalance: An Empirical Study

Abstract: Class imbalance has drawn much attention of researchers in software defect prediction. In practice, the performance of defect prediction models may be affected by the class imbalance problem. In this paper, we present an approach to evaluating the performance stability of defect prediction models on imbalanced datasets. First, random sampling is applied to convert the original imbalanced dataset into a set of new datasets with different levels of imbalance ratio. Second, typical prediction models are selected … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
31
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 34 publications
(32 citation statements)
references
References 37 publications
1
31
0
Order By: Relevance
“…To determine the performance stability of prediction models, Co-efficient of Variation (C.V) was applied to the results of the prediction models. C.V which is the percentage ratio of standard deviation (SD) and average (AVE) is used to remove the effect of average difference on the comparison stability [15,40]. The formula for C.V is given as thus:…”
Section: Accuracy =mentioning
confidence: 99%
See 1 more Smart Citation
“…To determine the performance stability of prediction models, Co-efficient of Variation (C.V) was applied to the results of the prediction models. C.V which is the percentage ratio of standard deviation (SD) and average (AVE) is used to remove the effect of average difference on the comparison stability [15,40]. The formula for C.V is given as thus:…”
Section: Accuracy =mentioning
confidence: 99%
“…SDP can be regarded as a classification task that involves categorizing software modules either as defective or non-defective, based on historical data and software metrics or features [14][15][16]. Software features or metrics reflect the characteristics of software modules.…”
Section: Introductionmentioning
confidence: 99%
“…As well as comparing the class correlation values and classification performance values, we conduct Wilcoxon matched-pair signed-rank test [50] with 95% confidence interval, a nonparametric test method for two or more related samples, to test whether the difference of class correlation values among five process metrics is significant and whether the classification performance difference of five process metrics is significant. This statistical method has been widely used in SDP [51], [52]. The original assumption is that there is no significant difference among five process metrics when the confidence interval is 95%.…”
Section: Gr(a) = Ig(s|a) Splite(a)mentioning
confidence: 99%
“…The data pre-processing is valuable to enhance the classification performance and decrease the time cost [44]- [46], which includes feature reduction and resampling techniques. Feature reduction is used to increase the generalization performance of classification [15], [47]- [53] by removing the irrelevant features from the balanced and imbalanced datasets. However, all these methods are focused on binary imbalance problem.…”
Section: Related Workmentioning
confidence: 99%