2018
DOI: 10.1002/stvr.1658
|View full text |Cite
|
Sign up to set email alerts
|

Heterogeneous fault prediction with cost‐sensitive domain adaptation

Abstract: Summary In the early phases of software testing, projects may have only limited historical defect data. Learning prediction model with such insufficient training data will limit the efficacy of learned predictor. In practice, there are usually many publicly available fault prediction datasets. Recently, heterogeneous fault prediction (HFP) has been proposed. However, existing HFP models do not investigate how to use mixed project data to predict target. Furthermore, defect data are often imbalanced. The imbala… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
10
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 23 publications
(10 citation statements)
references
References 75 publications
0
10
0
Order By: Relevance
“…The within-project defect prediction model to detect the DP instances is constructed from historical defect data of the same project [30,46]. However, in practice, there is not a sufficient amount of such historical data for some projects [47,48], so cross-project defect prediction approach is adopted to predict defects in a project via prediction models trained from historical defect data of other projects [11,[49][50][51]. This article is based on the within-project defect prediction settings.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…The within-project defect prediction model to detect the DP instances is constructed from historical defect data of the same project [30,46]. However, in practice, there is not a sufficient amount of such historical data for some projects [47,48], so cross-project defect prediction approach is adopted to predict defects in a project via prediction models trained from historical defect data of other projects [11,[49][50][51]. This article is based on the within-project defect prediction settings.…”
Section: Related Workmentioning
confidence: 99%
“…It can be beneficial to guide efforts of software testing and can be effective in optimal allocation of testing resources [5][6][7][8]. Yet, one of the challenges of SDP is CIP [9][10][11]. CIP means that the number of NDP artifacts is much more than the artifacts that have defects.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Li et al proposed a new cost sensitive transfer kernel canonical correlation analysis (CTKCCA) approach for HDP, which made the data distributions of source and target projects much more similar in the nonlinear feature space [3]. Li et al not only made better use of two projects but also alleviated the class imbalance problem by setting different misclassification costs for different samples [4]. Li et al proposed a multi-source selection based manifold discriminant alignment (MSMDA) approach.…”
Section: Introductionmentioning
confidence: 99%
“…Software defect prediction (SDP) is a hot research topic in current software engineering research domain, and it can help to optimize test resource allocation by predicting defect‐prone modules (the granularity of the module can be set as file, class, code change as needed) in advance . A number of defect prediction approaches have been proposed, and these approaches mainly apply machine learning techniques to build prediction models by mining data stored in software historical repositories . These approaches typically use various features (ie, metrics) to measure extracted modules from repositories and then apply machine learning algorithms to predict if a new module is defective or not.…”
Section: Introductionmentioning
confidence: 99%