2019
DOI: 10.1109/jxcdc.2019.2923745
|View full text |Cite
|
Sign up to set email alerts
|

A Ferroelectric FET-Based Processing-in-Memory Architecture for DNN Acceleration

Abstract: This paper presents a ferroelectric FET (FeFET)-based processing-in-memory (PIM) architecture to accelerate the inference of deep neural networks (DNNs). We propose a digital in-memory vector-matrix multiplication (VMM) engine design utilizing the FeFET crossbar to enable bit-parallel computation and eliminate analog-to-digital conversion in prior mixed-signal PIM designs. A dedicated hierarchical network-on-chip (H-NoC) is developed for input broadcasting and on-the-fly partial results processing, reducing th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 53 publications
(33 citation statements)
references
References 27 publications
0
32
0
Order By: Relevance
“…When the result of sense amplifier for CBL is 1, the CBL peripheral sends the partial sum, 1, to the H-NoC router and pre-discharges the CBL. The H-NoC connects the synaptic arrays, accumulates the partial sums, and sends the VMM result to the neuron module (Long et al, 2019 ).…”
Section: Hardware Architecturementioning
confidence: 99%
“…When the result of sense amplifier for CBL is 1, the CBL peripheral sends the partial sum, 1, to the H-NoC router and pre-discharges the CBL. The H-NoC connects the synaptic arrays, accumulates the partial sums, and sends the VMM result to the neuron module (Long et al, 2019 ).…”
Section: Hardware Architecturementioning
confidence: 99%
“…al. [19] and the SIGMA [26] accelerator for sparse computation. A detail design of the hybrid PIM is beyond the scope of this paper.…”
Section: Architectural Considerationsmentioning
confidence: 99%
“…Figure 10 shows the layer-wise distribution of protected and total computations. The throughput and efficiency are estimated considering the PIM [19] and sparse convolution accelerator modules [26], and the layer-wise OPs distribution ( Table 2). In ResNet18, 1% protected parameters translates to a 4% computational overhead.…”
Section: Architectural Considerationsmentioning
confidence: 99%
See 2 more Smart Citations