Proceedings of the 2020 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays 2020
DOI: 10.1145/3373087.3375306
|View full text |Cite
|
Sign up to set email alerts
|

AutoDNNchip

Abstract: Recent breakthroughs in Deep Neural Networks (DNNs) have fueled a growing demand for domain-specific hardware accelerators (i.e., DNN chips). However, designing DNN chips is non-trivial because: (1) mainstream DNNs have millions of parameters and operations; (2) the design space is large due to the numerous design choices of dataflows, processing elements, memory hierarchy, etc.; and (3) an algorithm/hardware co-design is needed to allow the same DNN functionality to have a different decomposition, which would… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 51 publications
(3 citation statements)
references
References 33 publications
0
3
0
Order By: Relevance
“…The definitions specify simulation functionality, execution costs like area, latency, or energy [46], or hardware synthesis of each component [43]. Frequently used high-level components can also be defined, such as vector units, systolic array, and computation/memory tiles [47], as specified in libraries like MAGNet [1] and AutoDNNChip [48]. Library also integrates reliability and security costs for components and specialized components such as razor flip flops for detecting timing violations or trusted memory.…”
Section: B End-to-end Agile Design Workflowmentioning
confidence: 99%
“…The definitions specify simulation functionality, execution costs like area, latency, or energy [46], or hardware synthesis of each component [43]. Frequently used high-level components can also be defined, such as vector units, systolic array, and computation/memory tiles [47], as specified in libraries like MAGNet [1] and AutoDNNChip [48]. Library also integrates reliability and security costs for components and specialized components such as razor flip flops for detecting timing violations or trusted memory.…”
Section: B End-to-end Agile Design Workflowmentioning
confidence: 99%
“…Timeloop [3], MAESTRO [4], and DeepOpt [5] propose DNN dataflow analysis frameworks for inference accelerators, but their evaluations focus only on convolution layers, and the work in [6] proposes an energy estimation model for the convolution operation only. While the modeling efforts in [7], [8] include pooling and tensor addition along with convolution, their scope is limited to inference and do not have support for training operations. TRIM [9] proposes a design space explorer for DNN training, but does not support batch normalization, a heavy training workload, and is not evaluated on mainstream ASIC accelerators [1], [10]- [13] that use systolic or vector dot-product style hardware.…”
Section: Introductionmentioning
confidence: 99%
“…However, because the conventional design flow of RTL programming is complicated and error-prone, it takes considerable design efforts to realize the customized GNN accelerator case by case for the target GNN workload or on the target FPGA devices. Though there is a lot of prior study on automating the flow of DNN accelerator development for FPGA [5,24,31], it is non-trivial to directly apply prior DNN-FPGA automation frameworks to the GNN-based applications. The reason is mainly attributed to three factors including 1) they are built on the traditional deep learning frameworks such as Caffe and Pytorch.…”
Section: Introductionmentioning
confidence: 99%