NAFIPS 2007 - 2007 Annual Meeting of the North American Fuzzy Information Processing Society 2007
DOI: 10.1109/nafips.2007.383813
|View full text |Cite
|
Sign up to set email alerts
|

Applying Novel Resampling Strategies To Software Defect Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
66
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 129 publications
(70 citation statements)
references
References 16 publications
0
66
0
Order By: Relevance
“…As a result, except for dynamic features (i.e., the features collected through executing the original program), we also adopt some static code information as features to indicate dynamic mutant execution information. Such information could complement dynamic features for precise prediction of software engineering problems, which is also replete in the literature [37], [38], [39]. Actually, the entire static analysis area in the end aims to guarantee that dynamic program executions satisfy certain properties.…”
Section: Identified Featuresmentioning
confidence: 99%
“…As a result, except for dynamic features (i.e., the features collected through executing the original program), we also adopt some static code information as features to indicate dynamic mutant execution information. Such information could complement dynamic features for precise prediction of software engineering problems, which is also replete in the literature [37], [38], [39]. Actually, the entire static analysis area in the end aims to guarantee that dynamic program executions satisfy certain properties.…”
Section: Identified Featuresmentioning
confidence: 99%
“…Resampling is an effective way to mitigate the effects of imbalanced data in change classification [18], [38]. Different algorithms are used to change the distribution between the majority class and the minority class.…”
Section: Resamplingmentioning
confidence: 99%
“…When the buggy rate is low, it is challenging to learn accurate models because there are fewer positive instances (i.e., buggy changes) for learning. Classifying imbalanced data is a known open challenge [18].…”
Section: Introductionmentioning
confidence: 99%
“…Since the standard classifiers are not applicable for imbalanced data, in order to deal with the problem of imbalanced data in this study, SMOTE was employed. This technique was proposed by Chawla et al [3] which is a famous re-sampling method in data pre-processing and has been applied in several articles, such as Pelayo and Dick [4], Zhao et al [5], Gu et al [6]. Using SMOTE technique, the number of samples in minority class can be increased by creating new synthetic samples instead of repeating them, so that the over-fitting problem in learning algorithm is avoided.…”
Section: Introductionmentioning
confidence: 99%