2019
DOI: 10.1109/jssc.2018.2880918
|View full text |Cite
|
Sign up to set email alerts
|

CONV-SRAM: An Energy-Efficient SRAM With In-Memory Dot-Product Computation for Low-Power Convolutional Neural Networks

Abstract: This work presents an energy-efficient SRAM with embedded dot-product computation capability, for binary-weight convolutional neural networks. A 10T bit-cell based SRAM array is used to store the 1-b filter weights. The array implements dotproduct as a weighted average of the bit-line voltages, which are proportional to the digital input values. Local integrating ADCs compute the digital convolution outputs, corresponding to each filter. We have successfully demonstrated functionality (> 98% accuracy) with the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
142
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 286 publications
(156 citation statements)
references
References 22 publications
0
142
0
Order By: Relevance
“…Moreover, the assumption is made that the ADC sampling time for an 8-bit output resolution is below the digital circuits' cycle time [23] so that n cyc,adc = 1. Note that (18) gives evidence for the linear relation between latency and the number of weight and input bits.…”
Section: Analog Accumulationmentioning
confidence: 99%
See 2 more Smart Citations
“…Moreover, the assumption is made that the ADC sampling time for an 8-bit output resolution is below the digital circuits' cycle time [23] so that n cyc,adc = 1. Note that (18) gives evidence for the linear relation between latency and the number of weight and input bits.…”
Section: Analog Accumulationmentioning
confidence: 99%
“…where n cyc was defined in (18). Consequently, the local read procedures also enter the energy balance as an array-sizedependent term E read,local .…”
Section: A Imc Subblocksmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, it also has great advantages, because the best feature extraction classification is automatically obtained from the training dataset. Start with LeNet-5 [24,25], a CNN particularly possesses normative structure-stacked convolutional layers (randomly followed by contrast max-pooling and normalization) followed by at least a completely linked layer. By comparing with shallow learning frameworks as well as handcrafted characteristics, the CNN architecture requires less knowledge of the domain.…”
Section: Related Workmentioning
confidence: 99%
“…Several papers have proven the capability of analog computing elements to achieve an acceptable trade-off between algorithmic accuracy and numerical precision. Analog solutions are also suitable to be implemented with an in-memory circuit architecture [17], [18], avoiding costly memory access.…”
Section: Introductionmentioning
confidence: 99%