Proceedings of the Fifteenth International Conference on Predictive Models and Data Analytics in Software Engineering 2019
DOI: 10.1145/3345629.3351449
|View full text |Cite
|
Sign up to set email alerts
|

Does chronology matter in JIT defect prediction?

Abstract: BACKGROUND: Just-In-Time (JIT) models, unlike the traditional defect prediction models, detect the fix-inducing changes (or defect inducing changes). These models are designed based on the assumption that past code change properties are similar to future ones. However, as the system evolves, the expertise of developers and/or the complexity of the system also change.AIM: In this work, we aim to investigate the effect of code change properties on JIT models over time. We also study the impact of using recent da… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 30 publications
0
4
0
Order By: Relevance
“…Machine learning in JIT-SDP looks at a wide range of classifier techniques, from standalone learners to ensemblebased learners, to find the best models. Commonly used standalone classifiers include Logistic Regression [3,23,24], Naïve bayes [25], Support Vector Machine [26], Decision Tree [27], and Neural Network [28]. Whereas for ensemblebased learners range from single ensemble learner such as Random Forest [29] to multi-layer ensembles [27,30,31].…”
Section: A Modelling Of Software Defect Predictionmentioning
confidence: 99%
“…Machine learning in JIT-SDP looks at a wide range of classifier techniques, from standalone learners to ensemblebased learners, to find the best models. Commonly used standalone classifiers include Logistic Regression [3,23,24], Naïve bayes [25], Support Vector Machine [26], Decision Tree [27], and Neural Network [28]. Whereas for ensemblebased learners range from single ensemble learner such as Random Forest [29] to multi-layer ensembles [27,30,31].…”
Section: A Modelling Of Software Defect Predictionmentioning
confidence: 99%
“…As a result, this method is good for both computational efficiency and interpretation. The Breakdown library [36] is a machine learning library that can be used to analyze the contributions of individual variables in a predictive model's output. The library can be used to construct both binary classifiers and regression models.…”
Section: Break-downmentioning
confidence: 99%
“…The authors of [36] investigated whether the chronological order of data used in just-in-time (JIT) defect prediction models affects their performance. The authors partially replicated a previous study and evaluated the impact of different training and testing data sets on the performance of the model.…”
Section: Supervised Learning Techniquesmentioning
confidence: 99%
“…Our dataset covers diverse domains and programming languages. Note that the datasets used in the previous studies [4], [14], [9], [15] do not contain our new metrics and cannot be directly used for our experiments.…”
Section: B Datasetmentioning
confidence: 99%