2020
DOI: 10.48550/arxiv.2002.04116
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Co-Exploration of Neural Architectures and Heterogeneous ASIC Accelerator Designs Targeting Multiple Tasks

Abstract: Neural Architecture Search (NAS) has demonstrated its power on various AI accelerating platforms such as Field Programmable Gate Arrays (FPGAs) and Graphic Processing Units (GPUs). However, it remains an open problem how to integrate NAS with Application-Specific Integrated Circuits (ASICs), despite them being the most powerful AI accelerating platforms. The major bottleneck comes from the large design freedom associated with ASIC designs. Moreover, with the consideration that multiple DNNs will run in paralle… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
1

Relationship

3
3

Authors

Journals

citations
Cited by 8 publications
(16 citation statements)
references
References 23 publications
0
16
0
Order By: Relevance
“…Neural Architecture Search (NAS) has achieved state-of-the-art performance in various perceptual tasks, such as image classifications [22,23], inference security [2] and image segmentation [20].…”
Section: Neural Architecture Searchmentioning
confidence: 99%
“…Neural Architecture Search (NAS) has achieved state-of-the-art performance in various perceptual tasks, such as image classifications [22,23], inference security [2] and image segmentation [20].…”
Section: Neural Architecture Searchmentioning
confidence: 99%
“…al, [59] uses Bayesian optimization, RELEASE [7] uses RL, ATLAS [84] uses black box optimizations, some compiler design [12], [50] use profile-guided optimization to perform target-independent front-end compiler optimizations on DNNs or linear algebra computations. Some recent works use RL on HW/SW co-exploration to explore both DNN and its mapping over hardware [6], [32], [44], [88]. The problem of mapping the DNN computation graph over multiple devices (CPU/GPU/TPU [34]) has also been explored through manual heuristics [8], [72], [91] and RL [24], [46], [51].…”
Section: Related Workmentioning
confidence: 99%
“…DNN Algorithm and Accelerator Co-exploration. Exploring the networks and the corresponding accelerators in a joint manner [1,31,39,40,50,92] has shown great potential towards maximizing both accuracy and efficiency. Recent works have extended NAS to jointly search DNN accelerators in addition to DNN structures.…”
Section: Related Workmentioning
confidence: 99%
“…Recent works have extended NAS to jointly search DNN accelerators in addition to DNN structures. In particular, [1,31,40,92] conducted RL-based searches to co-explore the network structures and design parameters of an FPGA-/ASIC-based accelerator, but their RL-based methods can suffer from large search costs, limiting their scalability to handle large joint spaces. Recently, [19,50] extended differentiable NAS to network and accelerator co-search.…”
Section: Related Workmentioning
confidence: 99%