Proceedings of the 28th ACM SIGPLAN Annual Symposium on Principles and Practice of Parallel Programming 2023
DOI: 10.1145/3572848.3577475
|View full text |Cite
|
Sign up to set email alerts
|

High-Performance GPU-to-CPU Transpilation and Optimization via High-Level Parallel Constructs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 48 publications
0
1
0
Order By: Relevance
“…The pre-pruning method cannot grasp the global information of the tree and has certain blindness, and this method involves difficulty to determine whether the child nodes of clipped nodes have existence value, which may cause the decision tree to stop growing prematurely, and thus it cannot obtain the optimal decision tree. In addition, the post-pruning algorithm is usually based on statistical knowledge, and some of its parameters either depend on prior statistical laws, the domain knowledge of experts, or certain assumptions, and often need to go through repeated comparative tests to obtain satisfactory results [44].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The pre-pruning method cannot grasp the global information of the tree and has certain blindness, and this method involves difficulty to determine whether the child nodes of clipped nodes have existence value, which may cause the decision tree to stop growing prematurely, and thus it cannot obtain the optimal decision tree. In addition, the post-pruning algorithm is usually based on statistical knowledge, and some of its parameters either depend on prior statistical laws, the domain knowledge of experts, or certain assumptions, and often need to go through repeated comparative tests to obtain satisfactory results [44].…”
Section: Methodsmentioning
confidence: 99%
“…prematurely, and thus it cannot obtain the optimal decision tree. In addition, the postpruning algorithm is usually based on statistical knowledge, and some of its parameters either depend on prior statistical laws, the domain knowledge of experts, or certain assumptions, and often need to go through repeated comparative tests to obtain satisfactory results [44].…”
Section: Methodsmentioning
confidence: 99%
“…It should be noted that OP2 is distinct from a software architecture design that uses transpilers for back-end development, which focuses on facilitating the development of back-end layers using transpilers and language interoperability for shared codebase development, whereas OP2 provides source-to-source translation and compilation framework for execution on different hardware platforms and is specifically designed for unstructured meshbased applications. • High-Performance GPU-to-CPU Transpilation and Optimization via High-Level Parallel Constructs [54] This article proposes an automated method to convert programs written in one programming model, such as CUDA, to another model, such as CPU threads, using Polygeist/MLIR, which can achieve performance portability and eliminate the need for manual application portability. This framework allows compiler transformations to be applied seamlessly, including parallelism-specific optimizations.…”
mentioning
confidence: 99%