2023
DOI: 10.1145/3555807
|View full text |Cite
|
Sign up to set email alerts
|

QUIDAM: A Framework for Qu ant i zation-aware D NN A ccelerator and M odel Co-Exploration

Abstract: As the machine learning and systems communities strive to achieve higher energy-efficiency through custom deep neural network (DNN) accelerators, varied precision or quantization levels, and model compression techniques, there is a need for design space exploration frameworks that incorporate quantization-aware processing elements into the accelerator design space while having accurate and fast power, performance, and area models. In this work, we present QUIDAM , a highly parameterized… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 36 publications
0
2
0
Order By: Relevance
“…Several prior works(e.g., [22,51,55,128]) investigate the new requirements, tradeoffs, and challenges that arise from using the CIM paradigm (e.g., dealing with non-idealities in the analog operations). To our knowledge, no work has proposed a complete solution or framework for these challenges; thus, this area requires further investigation.…”
Section: Computation-in-memory Acceleratorsmentioning
confidence: 99%
“…Several prior works(e.g., [22,51,55,128]) investigate the new requirements, tradeoffs, and challenges that arise from using the CIM paradigm (e.g., dealing with non-idealities in the analog operations). To our knowledge, no work has proposed a complete solution or framework for these challenges; thus, this area requires further investigation.…”
Section: Computation-in-memory Acceleratorsmentioning
confidence: 99%
“…Workloads worsen the memory bottleneck that lowers the overall performance of systems where deep learning models are run since neural network workloads continue to have big memory footprints and significant computational requirements to attain improved accuracy (Inci 2022;Inci et al, 2022a). As deep learning models have become more proficient in many tasks, developing smaller neural network models in terms of trainable parameters has attracted interest from many researchers in the field (Inci et al, 2022b) to support operational needs in flood forecasting (Krajewski et al, 2021;Xiang and Demir, 2022) and inundation mapping (Hu and Demir, 2021;Li et al, 2022).…”
Section: Introductionmentioning
confidence: 99%