2019
DOI: 10.1016/j.infsof.2019.04.005
|View full text |Cite
|
Sign up to set email alerts
|

“Bad smells” in software analytics papers

Abstract: CONTEXT: There has been a rapid growth in the use of data analytics to underpin evidence-based software engineering. However the combination of complex techniques, diverse reporting standards and poorly understood underlying phenomena are causing some concern as to the reliability of studies. OBJECTIVE: Our goal is to provide guidance for producers and consumers of software analytics studies (computational experiments and correlation studies). METHOD: We propose using "bad smells", i.e., surface indications of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
24
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 29 publications
(24 citation statements)
references
References 103 publications
(140 reference statements)
0
24
0
Order By: Relevance
“…The predictive accuracy of the defect prediction model heavily relies on the modelling pipelines of defect prediction models [4,22,56,73,74,76,78]. To accurately predicting defective areas of code, prior studies conducted a comprehensive evaluation to identify the best technique of the modelling pipelines for defect models.…”
Section: The Modelling Pipeline Of Defect Prediction Modelsmentioning
confidence: 99%
“…The predictive accuracy of the defect prediction model heavily relies on the modelling pipelines of defect prediction models [4,22,56,73,74,76,78]. To accurately predicting defective areas of code, prior studies conducted a comprehensive evaluation to identify the best technique of the modelling pipelines for defect models.…”
Section: The Modelling Pipeline Of Defect Prediction Modelsmentioning
confidence: 99%
“…6,14,15,[19][20][21][22][23] Recently, researchers have also highlighted that the lack of reference models can be considered as an additional cause for inconsistent findings, and for this reason, the adoption of benchmarking mechanisms is highly recommended as a good practice in decision making. 3,6,23,[33][34][35][42][43][44][45] The proposed reference models can be categorized into "worst-case scenarios" 6,23,33 and "reasonably well standards" 34,35 serving different purposes in the evaluation process. A worst-case model, (eg, the Median model 33 , the Mean model, 23 or the Random model 6 ) can be used in practice when the objective is to assess whether a newly proposed model can be considered as a promising candidate for the estimation task of forthcoming projects.…”
Section: Related Work and Contributionmentioning
confidence: 99%
“…The latter is an important requirement for any new candidate characterizing the quality of the derived solutions, since the inability of a proposed model to predict better than a naïve approach raises significant doubts about its practical value. 45 In other words, worst-case reference models contribute to a straightforward sanity-check scope, since they provide borderline standards that any proposed model should outperform in order to qualify as a potentially useful option. 35,45 Moreover, Menzies et al 14 point out that there is a necessity for introducing "pruning" mechanisms, since practitioners face the problem of being overwhelmed by a growing number of possibly useless SDEE methods.…”
Section: Related Work and Contributionmentioning
confidence: 99%
See 1 more Smart Citation
“…There has been rapid growth in the use of data analytics to support evidence-based software engineering [14,20]. Modern software development relies on short feedback cycles as a way to provide flexibility and rapid adaptation to market fluctuations.…”
Section: Introductionmentioning
confidence: 99%