Proceedings of the Ninth International Workshop on Dynamic Analysis 2012
DOI: 10.1145/2338966.2336798
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating program analysis and testing tools with the RUGRAT random benchmark application generator

Abstract: Benchmarks are heavily used in different areas of computer science to evaluate algorithms and tools. In program analysis and testing, open-source and commercial programs are routinely used as benchmarks to evaluate different aspects of algorithms and tools. Unfortunately, many of these programs are written by programmers who introduce different biases, not to mention that it is very difficult to find programs that can serve as benchmarks with high reproducibility of results.We propose a novel approach for gene… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2014
2014
2022
2022

Publication Types

Select...
3
1
1

Relationship

2
3

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 33 publications
0
3
0
Order By: Relevance
“…Threats to external validity limit the extent to which the results will generalize to other kinds of concurrent programs. Whereas there are large benchmark suites and benchmark generators for sequential programs, there is no large and widely accepted bug benchmark suites for concurrent programs, and thus this threat is common to all prior work in this area. To mitigate this problem, the benchmark suite includes most of the benchmarks used in other work and furthermore, includes both Java and C++ programs.…”
Section: Empirical Studiesmentioning
confidence: 99%
“…Threats to external validity limit the extent to which the results will generalize to other kinds of concurrent programs. Whereas there are large benchmark suites and benchmark generators for sequential programs, there is no large and widely accepted bug benchmark suites for concurrent programs, and thus this threat is common to all prior work in this area. To mitigate this problem, the benchmark suite includes most of the benchmarks used in other work and furthermore, includes both Java and C++ programs.…”
Section: Empirical Studiesmentioning
confidence: 99%
“…A practical challenge for these kinds of generators is to construct realistic programs. However, an empirical study indicates that it is statistically impossible for a program analysis technique to differentiate a program written by a human from one that the tool generates [20]. The study compared real and generated programs with 78 existing software metrics.…”
Section: Objects Of Analysismentioning
confidence: 99%
“… This article is an extended version of our previous work presented at the 10th International Workshop on Dynamic Analysis (WODA 2012) .…”
mentioning
confidence: 99%