2019
DOI: 10.1007/s11227-019-02880-z
|View full text |Cite
|
Sign up to set email alerts
|

A methodology correlating code optimizations with data memory accesses, execution time and energy consumption

Abstract: The advent of data proliferation and electronic devices gets low execution time and energy consumption software in the spotlight. The key to optimizing software is the correct choice, order as well as parameters of optimizations-transformations, that has remained an open problem in compilation research for decades for various reasons. First, most of the transformations are interdependent and thus addressing them separately is not effective. Second, it is very hard to couple the transformation parameters to the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 53 publications
0
3
0
Order By: Relevance
“…As a result of the parameters being designed by expert programmers, most of these search spaces are relatively small compared to some other autotuning works 12,13 -the number of tuning parameters is not needlessly excessive, and most parameters only have a few rationally-selected values to choose from. While such spaces are hard to search using methods that partially rely on local search (such as simulated annealing), because most of their dimensions are very small and discontinuous, they also contain a higher proportion of relatively fast configurations, which makes random search a viable strategy.…”
Section: Benchmarksmentioning
confidence: 99%
“…As a result of the parameters being designed by expert programmers, most of these search spaces are relatively small compared to some other autotuning works 12,13 -the number of tuning parameters is not needlessly excessive, and most parameters only have a few rationally-selected values to choose from. While such spaces are hard to search using methods that partially rely on local search (such as simulated annealing), because most of their dimensions are very small and discontinuous, they also contain a higher proportion of relatively fast configurations, which makes random search a viable strategy.…”
Section: Benchmarksmentioning
confidence: 99%
“…Due to the problem of finding the optimum tile size is very complex and includes a vast exploration space [9], in addition to general methods, a large number of algorithm-specific analytical models also exist for Matrix-Matrix Multiplication (MMM) [12] [14], Matrix-Vector Multiplication [13], tensor contractions [15], Fast Fourier Transform [10], stencil [23] and other algorithms, but the proposed approaches cannot be generalized. In particular, regarding stencil applications, there has been a long thread of research and development tackling data locality and parallelism, where many loop tiling strategies have been proposed such as overlapped tiling [26] [5], diamond tiling [2] and others.…”
Section: Related Workmentioning
confidence: 99%
“…Researchers have suggested using parallel programming techniques, including MPI, multiprocessing, and threading, to improve the efficiency of RYU and POX. These methods can better utilize the CPU's processing capabilities [6]- [8] to improve the performance of these two controllers [9] in terms of processing speed and memory allocation. The effectiveness of parallel methods in improving the efficiency of network controllers has not been thoroughly investigated.…”
Section: Introductionmentioning
confidence: 99%