Proceedings of the 32nd ACM/IEEE International Conference on Software Engineering - Volume 2 2010
DOI: 10.1145/1810295.1810313
|View full text |Cite
|
Sign up to set email alerts
|

Transparent combination of expert and measurement data for defect prediction

Abstract: Defining strategies on how to perform quality assurance (QA) and how to control such activities is a challenging task for organizations developing or maintaining software and softwareintensive systems. Planning and adjusting QA activities could benefit from accurate estimations of the expected defect content of relevant artifacts and the effectiveness of important quality assurance activities. Combining expert opinion with commonly available measurement data in a hybrid way promises to overcome the weaknesses … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2013
2013
2022
2022

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 27 publications
(17 citation statements)
references
References 24 publications
0
14
0
Order By: Relevance
“…One alternative would be 10-fold cross validation, but it essentially uses 90% training data and 10% test data. Moreover, random split is widely used in the literature [5], [9], [10].…”
Section: ) Within-project Predictionmentioning
confidence: 99%
See 1 more Smart Citation
“…One alternative would be 10-fold cross validation, but it essentially uses 90% training data and 10% test data. Moreover, random split is widely used in the literature [5], [9], [10].…”
Section: ) Within-project Predictionmentioning
confidence: 99%
“…Most approaches employ machine learning classifiers to build a prediction model from data sets mined from software repositories, and the model is used to identify software defects. However, most approaches are evaluated in within-project settings, i.e., a prediction model is built from a part of a project and the model is evaluated with the remainder of the project by 10-fold cross validation [2], [6], [7], [8] and/or random instance splits [5], [9], [10].…”
Section: Introductionmentioning
confidence: 99%
“…We use 50:50 random splits, which are widely used in the evaluation of defect prediction models [22,37,41]. For the 50:50 random splits, we use one half of the instances for training a model and the rest for test (round 1).…”
Section: Experimental Designmentioning
confidence: 99%
“…In our study, code characteristics are chosen independent of specific hardware description languages, which is listed in Table I. 2) History Characteristics: History information, such as past changes, fixes, bugs, and so on, may also have significant impacts on bug occurrence. In the domain of software engineering, history information has already been demonstrated to be helpful in predicting software defects [20]- [22], [26], [31], [42]. Hence, the proposed pre-silicon bug forecast framework also utilizes history information.…”
Section: ) Code Characteristicsmentioning
confidence: 99%
“…In the field of software engineering, many studies have been dedicated to characterize the relationship between the software characteristics and fault-proneness to assess the design quality [5], [19]- [22], [26], [27], [30], [31], [42], which mainly focused on selecting characteristics that have most impacts on the fault-proneness of software.…”
Section: B Defect Prediction In Software Engineeringmentioning
confidence: 99%