2020
DOI: 10.1109/jxcdc.2020.2992306
|View full text |Cite
|
Sign up to set email alerts
|

A Relaxed Quantization Training Method for Hardware Limitations of Resistive Random Access Memory (ReRAM)-Based Computing-in-Memory

Abstract: Nonvolatile computing-in-memory (nvCIM) exhibits high potential for neuromorphic computing involving massive parallel computations and for achieving high energy efficiency. nvCIM is especially suitable for deep neural networks, which are required to perform large amounts of matrix-vector multiplications. However, a comprehensive quantization algorithm has yet to be developed, which overcomes the hardware limitations of resistive random access memory (ReRAM)-based nvCIM, such as the number of I/O, word lines (W… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 10 publications
0
6
0
Order By: Relevance
“…Although the IMC technology achieves high energy efficiency, when DNNs trained in software are deployed on IMC hardware, accuracy degradation can occur due to limited ADC precision, variations in the IMC devices, ambient conditions, and transistor non-linearity [11,13,19,25,27,28]. Several recent works have attempted to address these particular issues, and representative NVM-based works are described below.…”
Section: Hardware-aware Dnn Training For Accurate Dnn Inference With ...mentioning
confidence: 99%
See 1 more Smart Citation
“…Although the IMC technology achieves high energy efficiency, when DNNs trained in software are deployed on IMC hardware, accuracy degradation can occur due to limited ADC precision, variations in the IMC devices, ambient conditions, and transistor non-linearity [11,13,19,25,27,28]. Several recent works have attempted to address these particular issues, and representative NVM-based works are described below.…”
Section: Hardware-aware Dnn Training For Accurate Dnn Inference With ...mentioning
confidence: 99%
“…However, additional area and energy overhead are incurred in the IMC hardware due to the addition of thermal reference cells. A quantization-aware DNN training scheme was proposed in [28] which considered input and weight quantization, RRAM-based convolution, and ADC quantization. However, only up to 36 rows are activated simultaneously for IMC to limit the accuracy degradation, and still, >2% accuracy loss is reported for the CIFAR-10 dataset.…”
Section: Hardware-aware Dnn Training For Accurate Dnn Inference With ...mentioning
confidence: 99%
“…Quantization is a discrete process, making it a suitable candidate to be learned using a Gumbel-based estimator. The authors of [61], [62] propose to learn the quantization levels, which form a quantization codebook, whereas the authors of [63] learn to select such a codebook as whole. Moreover, dataadaptive binarization has been learned as well [64], [65], [66].…”
Section: Data Compressionmentioning
confidence: 99%
“…While effective in reducing ADC resolution requirements, bitslicing results in severe area overhead. The work in [14] Tony Liu, Amirali Amirsoleimani, Jianxiong Xu and Roman Genov are with the Edward S. Rogers Sr. Department of Electrical and Computer Engineering, University of Toronto, 10 King's College Road, Toronto, Ontario, Canada. Fabien Alibart, Yann Beilliard, Serge Ecoffey, and Dominique Drouin are with the Department of Electrical and Computer Engineering, University of Sherbrooke, QC, Canada.…”
Section: Introductionmentioning
confidence: 99%