Proceedings of the 2023 ACM/SIGDA International Symposium on Field Programmable Gate Arrays 2023
DOI: 10.1145/3543622.3573210
|View full text |Cite
|
Sign up to set email alerts
|

CHARM: C omposing H eterogeneous A ccele R ators for M atrix Multiply on Versal ACAP Architecture

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(8 citation statements)
references
References 22 publications
0
8
0
Order By: Relevance
“…This happens, e.g., when the FP32 output from Softmax needs to be used as the input of the next matrix-multiply layer. We deploy the same INT8 quantized model, DeiT-T, on the AMD ACAP architecture [30] VCK190 [27] board using CHARM [19]. CHARM [19] is the state-of-the-art deep learning inference accelerator and mapping framework on ACAP architecture, which features FPGA, AIE vector processors, and CPU on the system-on-chip.…”
Section: Design Challenges and Proposed Solutionmentioning
confidence: 99%
See 4 more Smart Citations
“…This happens, e.g., when the FP32 output from Softmax needs to be used as the input of the next matrix-multiply layer. We deploy the same INT8 quantized model, DeiT-T, on the AMD ACAP architecture [30] VCK190 [27] board using CHARM [19]. CHARM [19] is the state-of-the-art deep learning inference accelerator and mapping framework on ACAP architecture, which features FPGA, AIE vector processors, and CPU on the system-on-chip.…”
Section: Design Challenges and Proposed Solutionmentioning
confidence: 99%
“…However, it assumes a very flexible Network-on-Chip (NoC) to connect the accelerators which consumes non-negligible resources and may cause large overhead because of the data congestion in the NoC. CHARM [19] composes heterogeneous accelerators for deep learning applications on ACAP. However, CHARM does not support on-chip data forwarding which results in longer inference latency.…”
Section: Hybrid Acceleratorsmentioning
confidence: 99%
See 3 more Smart Citations