2017
DOI: 10.1007/s10703-016-0264-5
|View full text |Cite
|
Sign up to set email alerts
|

Empirical software metrics for benchmarking of verification tools

Abstract: We study empirical metrics for software source code, which can predict the performance of verification tools on specific types of software. Our metrics comprise variable usage patterns, loop patterns, as well as indicators of control-flow complexity and are extracted by simple data-flow analyses. We demonstrate that our metrics are powerful enough to devise a machine-learning based portfolio solver for software verification. We show that this portfolio solver would be the (hypothetical) overall winner of the i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
30
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(30 citation statements)
references
References 23 publications
0
30
0
Order By: Relevance
“…The advantage of using a diverse set of models is that we can identify the most suitable application areas. Furthermore, we compare lower lever parameters of CEGAR as opposed to most experiments in the literature [11,19,36,37], where different algorithms or tools are compared. We formulate and address a research question related to the effectiveness and efficiency of each of our contributions.…”
Section: Experimental Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…The advantage of using a diverse set of models is that we can identify the most suitable application areas. Furthermore, we compare lower lever parameters of CEGAR as opposed to most experiments in the literature [11,19,36,37], where different algorithms or tools are compared. We formulate and address a research question related to the effectiveness and efficiency of each of our contributions.…”
Section: Experimental Evaluationmentioning
confidence: 99%
“…Experimental evaluation There are many works in the literature that focus on experimental evaluation and comparison of model checking algorithms [11,19,36,37]. However, they usually focus on a certain domain (e.g., SV-COMP).…”
Section: Multiple Refinements For a Counterexamplementioning
confidence: 99%
“…The metrics to classify programs proposed in [13] are related to three different aspects: variable roles, loop patterns, and control flow.…”
Section: Program Metricsmentioning
confidence: 99%
“…In Section 2 we summarize the program metrics exploited in [13]. We propose corresponding metrics for term rewrite systems in Section 3.…”
Section: Introductionmentioning
confidence: 99%
“…The reason for this is that the analysis time of a SAT or SMT problem can vary significantly between two problems of the same size or between two solvers, due to the nature of modern solver algorithms [15]. For instance, compare adpcmencode (program size 4, 911 steps, timeout after 1 minute) with ndes (program size 5, 727 steps, solved in less than one second).…”
Section: Reduction Of Computational Effortmentioning
confidence: 99%