2016
DOI: 10.1002/cpe.4029
|View full text |Cite
|
Sign up to set email alerts
|

Machine learning‐based auto‐tuning for enhanced performance portability of OpenCL applications

Abstract: Summary Heterogeneous computing, combining devices with different architectures such as CPUs and GPUs, is rising in popularity and promises increased performance combined with reduced energy consumption. OpenCL has been proposed as a standard for programming such systems and offers functional portability. However, it suffers from poor performance portability, because applications must be retuned for every new device. In this paper, we use machine learning‐based auto‐tuning to address this problem. Benchmarks a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
31
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 20 publications
(31 citation statements)
references
References 53 publications
0
31
0
Order By: Relevance
“…The higher error for the execution time is due to the huge variations in time for the same program for larger input values. These results obtained for occupancy and eligible warps are better than the mean error rate produced by Falch and Elster for Nvidia GPUs. For execution time, the mean error rate is 2% higher than the results obtained in the aforementioned work .…”
Section: Framework Validation and Evaluationmentioning
confidence: 59%
See 4 more Smart Citations
“…The higher error for the execution time is due to the huge variations in time for the same program for larger input values. These results obtained for occupancy and eligible warps are better than the mean error rate produced by Falch and Elster for Nvidia GPUs. For execution time, the mean error rate is 2% higher than the results obtained in the aforementioned work .…”
Section: Framework Validation and Evaluationmentioning
confidence: 59%
“…These results obtained for occupancy and eligible warps are better than the mean error rate produced by Falch and Elster for Nvidia GPUs. For execution time, the mean error rate is 2% higher than the results obtained in the aforementioned work . This is justifiable with the dataset being used in this work, which is highly diverse, whereas it is not so in the work by by Falch and Elster …”
Section: Framework Validation and Evaluationmentioning
confidence: 59%
See 3 more Smart Citations