Proceedings of the 2012 International Conference on Compilers, Architectures and Synthesis for Embedded Systems 2012
DOI: 10.1145/2380403.2380423
|View full text |Cite
|
Sign up to set email alerts
|

The RACECAR heuristic for automatic function specialization on multi-core heterogeneous systems

Abstract: Embedded systems increasingly combine multi-core processors and heterogeneous resources such as graphics-processing units and field-programmable gate arrays. However, significant application design complexity for such systems caused by parallel programming and device-specific challenges has often led to untapped performance potential. Application developers targeting such systems currently must determine how to parallelize computation, create different device-specialized implementations for each heterogeneous … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2014
2014
2016
2016

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 29 publications
0
3
0
Order By: Relevance
“…For instance, a different concept of elasticity named elastic computing has been introduced [8] which supports portable designs (the ability to run one function on multiple platforms) but augmented with a selection model which can choose the best available implementation based on application properties, and pre-compiled performance models. The authors extend their work in [9] which introduces a heuristic for generating parallel implementations for heterogeneous computers.…”
Section: Background and Related Workmentioning
confidence: 99%
“…For instance, a different concept of elasticity named elastic computing has been introduced [8] which supports portable designs (the ability to run one function on multiple platforms) but augmented with a selection model which can choose the best available implementation based on application properties, and pre-compiled performance models. The authors extend their work in [9] which introduces a heuristic for generating parallel implementations for heterogeneous computers.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Performance information on each function is made available by implementation performance graphs (IPGs) similar to those described in [31], although any form of performance prediction could potentially be used. IPGs contain the execution times for a function implementation at a variety of input sizes.…”
Section: Scopes 2014mentioning
confidence: 99%
“…Lastly, our implementation of the scheduler does not consider intra-function parallelization optimizations, such as parallelizing a single function across multiple CPU cores using divide-andconquer techniques (e.g., [31]). Such optimizations are compatible with our approach and will be merged with our framework in the future.…”
Section: Limitations and Future Workmentioning
confidence: 99%