2018 IEEE/ACS 15th International Conference on Computer Systems and Applications (AICCSA) 2018
DOI: 10.1109/aiccsa.2018.8612819
|View full text |Cite
|
Sign up to set email alerts
|

Using Machine Learning Techniques to Classify and Predict Static Code Analysis Tool Warnings

Abstract: This paper discusses our work on using software engineering metrics (i.e., source code metrics) to classify an error message generated by a Static Code Analysis (SCA) tool as a true-positive, false-positive, or false-negative. Specifically, we compare the performance of Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Random Forests, and Repeated Incremental Pruning to Produce Error Reduction (RIPPER) over eight datasets. The performance of the techniques is assessed by computing the F-measure metric, w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(9 citation statements)
references
References 14 publications
0
9
0
Order By: Relevance
“…Benchmarks: We used C projects in CoreBench [14] (we used the first listed buggy version of all 4 projects) and BugBench [48] (it released 11 programs and 8 are C programs which we used). The 12 projects consist of 2.9 million lines of code (sloc) 7 , shown in the first two columns of Table 1. From the Commercial Tool and PolySpace, we processed a total of 1955 warnings of 41 categories.…”
Section: Static Analysis Toolsmentioning
confidence: 99%
See 1 more Smart Citation
“…Benchmarks: We used C projects in CoreBench [14] (we used the first listed buggy version of all 4 projects) and BugBench [48] (it released 11 programs and 8 are C programs which we used). The 12 projects consist of 2.9 million lines of code (sloc) 7 , shown in the first two columns of Table 1. From the Commercial Tool and PolySpace, we processed a total of 1955 warnings of 41 categories.…”
Section: Static Analysis Toolsmentioning
confidence: 99%
“…There are also approaches that identify patterns from warnings, source code and software repositories for predicting false positives [7,9,17,17,40,45,59,71,74], and that use machine learning techniques to learn what are likely true and false positives [7, 23, 39, 45, 59, 70? ]. For example, Zhang et al automatically learned and integrated the users' feedback to rank the warnings [76].…”
Section: Related Workmentioning
confidence: 99%
“…Then, they tagged the samples. Similarly, Alikhashashneh et al [11] used the Understand tool to detect various metrics, and employed them on the Juliet test suite for C++.…”
Section: Quality Assessment/predictionmentioning
confidence: 99%
“…Ribeiro et al [259] generated features only from the warnings (such as redundancy level and number of warnings in the same file). Some studies [11,158] used source code metrics as features.…”
Section: Quality Assessment/predictionmentioning
confidence: 99%
“…They proposed to selectively learn their SVM model by only harmless codeset structures used to predict only FPAs. In [21] and [22] authors uses ML techniques to reduce false Authors in [23] use lexical tokenization labeled by the human to learn their CNN classifier to reduce false alerts. They propose a continuous mechanism for code integration after review.…”
Section: A Machine Learning-based Approachesmentioning
confidence: 99%