2012
DOI: 10.1145/2189750.2150983
|View full text |Cite
|
Sign up to set email alerts
|

Iterative optimization for the data center

Abstract: Iterative optimization is a simple but powerful approach that searches for the best possible combination of compiler optimizations for a given workload. However, each program, if not each data set, potentially favors a different combination. As a result, iterative optimization is plagued by several practical issues that prevent it from being widely used in practice: a large number of runs are required for finding the best combination; the process can be data set dependent; and the exploration process incurs si… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
8
0

Year Published

2013
2013
2016
2016

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(8 citation statements)
references
References 27 publications
0
8
0
Order By: Relevance
“…(3) How does the strategy perform when datasets are nonuniformly mixed? We briefly answer them here, and more details can be found in the original conference paper [Chen et al 2012]. For the first question, we find that steady-state performance of IODC is within 98% of the best possible performance, in large part because IODC can keep the overhead of recompilations and training runs small.…”
Section: Discussionmentioning
confidence: 92%
See 2 more Smart Citations
“…(3) How does the strategy perform when datasets are nonuniformly mixed? We briefly answer them here, and more details can be found in the original conference paper [Chen et al 2012]. For the first question, we find that steady-state performance of IODC is within 98% of the best possible performance, in large part because IODC can keep the overhead of recompilations and training runs small.…”
Section: Discussionmentioning
confidence: 92%
“…The initialization phase does not take that long; after a couple thousand runs (2,670 runs on average), IODC yields a net performance benefit; the reason is that combinations that outperform the default -O3 production combination can usually be found quickly. We show the cumulated speedup profiles for six out of nine benchmarks (including three newly added benchmarks over the conference paper) in Figure 10; the other three benchmarks are shown in the conference paper [Chen et al 2012]. Overall, the proposed strategy achieves a cumulated speedup of 1.39× for bzip2e, 1.12× for blackscholes and freqmine, 1.10× for x264, 1.09× for kmeans and vips, 1.08× for canneal, 1.07× for streamcluster, and 1.06× for ferret, with an average of 1.12×.…”
Section: Benchmarkmentioning
confidence: 99%
See 1 more Smart Citation
“…In the dynamic autotuning space, there have been a number of systems developed [10,12,14,15,24,26,27,31,40] that focus on creating applications that can monitor and automatically tune themselves to optimize a particular objective. Many of these systems employ a control systems based autotuner that operates on a linear model of the application being tuned.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, they contain complex sharing patterns and interthread communications. Our evaluations on cloud applications are based on the Sector-Sphere platform [Gu et al 2009], with the datasets from [Chen et al 2012]. The two mapreduce applications are knn [Gillick et al 2006] and ann [Liu et al 2010].…”
Section: Benchmarksmentioning
confidence: 99%