2020 IEEE International Symposium on High Performance Computer Architecture (HPCA) 2020
DOI: 10.1109/hpca47549.2020.00015
|View full text |Cite
|
Sign up to set email alerts
|

SIGMA: A Sparse and Irregular GEMM Accelerator with Flexible Interconnects for DNN Training

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
167
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 307 publications
(207 citation statements)
references
References 30 publications
0
167
0
Order By: Relevance
“…There has been an incredible amount of interest in DNN hardware acceleration. Broadly speaking, the architecture community has focused on designing efficient dataflows to maximize local reuse of data and functional unit utilization [4,10,11,15,28,34,37,39], explore the space of possible dataflows and mappings [26,45,74], exploit model sparsity and data quantization [17,21,29,38,46,53,71,73,78], map DNN accelerators to FPGAs [20,66,69], and explore alternative compute, memory, and packaging technologies [35,58,59,67]. All of these works are highly relevant to this field.…”
Section: Related Workmentioning
confidence: 99%
“…There has been an incredible amount of interest in DNN hardware acceleration. Broadly speaking, the architecture community has focused on designing efficient dataflows to maximize local reuse of data and functional unit utilization [4,10,11,15,28,34,37,39], explore the space of possible dataflows and mappings [26,45,74], exploit model sparsity and data quantization [17,21,29,38,46,53,71,73,78], map DNN accelerators to FPGAs [20,66,69], and explore alternative compute, memory, and packaging technologies [35,58,59,67]. All of these works are highly relevant to this field.…”
Section: Related Workmentioning
confidence: 99%
“…To address this challenge, a GEMM accelerator for DNN training called SIGMA [71] is proposed that can handle various irregular GEMMs dimension and different levels of sparsity, while maximizing the utilization of computing resources. The Flexible Dot Product Engine (Flex-DPE) maps GEMMs of various dimension and sparsity levels to PEs using scalable interconnects.…”
Section: Reconfigurable Interconnectsmentioning
confidence: 99%
“…al. [19] and the SIGMA [26] accelerator for sparse computation. A detail design of the hybrid PIM is beyond the scope of this paper.…”
Section: Architectural Considerationsmentioning
confidence: 99%
“…In particular, our approach shows significant improvement over baselines, namely, device-variation-aware training [21] and gradient-based protection [12], for hardware friendly DNN models like MobileNet-V2. We evaluate overheads of Hessian-driven parameter protection considering existing PIM and digital accelerators for sparse convolution [19,26]. The analysis shows negligible throughput overhead, but 7.5%,19.5%, and 4.9% reduction in power-efficiency compared to the baseline PIM design [26] for ResNet, MobileNet, and DenseNet, respectively.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation