1994
DOI: 10.1145/190787.190807
|View full text |Cite
|
Sign up to set email alerts
|

The effect of compiler-flag tuning on SPEC benchmark performance

Abstract: The SPEC CINT92 and CFP92 benchmark suites are application-based system benchmarks primarily intended for workstation-class system performance measurements. The SPEC CPU benchmark results are widely disseminated by system vendors and as such have become the de-facto standard for comparing system performance. Recently, many observers have expressed concerns about the suitability of published SPEC benchmark results in representing application performance on typical systems. The most outspoken concern is that the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

1995
1995
2010
2010

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 6 publications
(4 reference statements)
0
3
0
Order By: Relevance
“…This makes the search for adequate flags a trial-and-error process. Moreover, as Chan et al pointed out in 1994 [10], it is difficult to ensure that none of the flags chosen has been added to the compiler just to optimize some SPEC program, a situation not allowed by any SPEC benchmark release. Besides this, the base speed metric forces to use the same flags to compile the entire benchmark.…”
Section: Compiling and Running Issuesmentioning
confidence: 99%
“…This makes the search for adequate flags a trial-and-error process. Moreover, as Chan et al pointed out in 1994 [10], it is difficult to ensure that none of the flags chosen has been added to the compiler just to optimize some SPEC program, a situation not allowed by any SPEC benchmark release. Besides this, the base speed metric forces to use the same flags to compile the entire benchmark.…”
Section: Compiling and Running Issuesmentioning
confidence: 99%
“…This soon prompted the question whether such excessive optimizations are representative for real programming, or whether it would be better to "burn all flags" [13]. The issue was later discussed in academic publications [1,8] and in the trade press. After a considerable debate, the final compromise for SPEC, established in January 1994, was that two metrics were established, "baseline" and "peak": Every SPEC CPU benchmark measurement has to measure performance with a restricted "baseline" set of flags (metric name: SPECint_base2000 or similar), and optionally with a more extended "peak" set of flags.…”
Section: Evolution Of the Spec Cpu Benchmarks Issuesmentioning
confidence: 99%
“…This makes the search for adequate flags a trial-and-error process. Moreover, as Chan et al pointed out in 1994 [32], it is difficult to ensure that none of the flags chosen is offered by the compiler just to optimize some SPEC program, a situation not allowed by any released SPEC benchmark. Besides this, the base speed metric forces to use the same flags to compile the entire benchmark.…”
Section: Compiler Flagsmentioning
confidence: 99%