2021
DOI: 10.1145/3460233
|View full text |Cite
|
Sign up to set email alerts
|

Impact of On-chip Interconnect on In-memory Acceleration of Deep Neural Networks

Abstract: With the widespread use of Deep Neural Networks (DNNs), machine learning algorithms have evolved in two diverse directions—one with ever-increasing connection density for better accuracy and the other with more compact sizing for energy efficiency. The increase in connection density increases on-chip data movement, which makes efficient on-chip communication a critical function of the DNN accelerator. The contribution of this work is threefold. First, we illustrate that the point-to-point (P2P)-based interconn… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
2

Relationship

2
6

Authors

Journals

citations
Cited by 14 publications
(4 citation statements)
references
References 39 publications
0
3
0
Order By: Relevance
“…The interface between NeuroSim and popular ML frameworks such as PyTorch and TensorFlow has also been created to make it more user-friendly [95]. However, one major drawback of NeuroSim is that it assumes H-Tree based bus interconnect for inter-tile communication, which can consume up to 90% of the total energy consumption of DNN inference [96]. To overcome this issue, Krishnan et al…”
Section: That Includes An Nop For On-package Communication Noc For On...mentioning
confidence: 99%
“…The interface between NeuroSim and popular ML frameworks such as PyTorch and TensorFlow has also been created to make it more user-friendly [95]. However, one major drawback of NeuroSim is that it assumes H-Tree based bus interconnect for inter-tile communication, which can consume up to 90% of the total energy consumption of DNN inference [96]. To overcome this issue, Krishnan et al…”
Section: That Includes An Nop For On-package Communication Noc For On...mentioning
confidence: 99%
“…Large NN layers are assigned to multiple PEs, and the partial sums are aggregated in a global buffer (Long et al, 2019;Krishnan et al, 2021). To maximize the processing parallelism of memristor crossbars, some memristor computing systems (Zhu et al, 2020;Wan et al, 2022) proposed to directly transfer the partial outputs of a NN layer to the PEs where its next layer is located.…”
Section: Preliminariesmentioning
confidence: 99%
“…The memory accessing patterns leak the NN structure information. For some memristor computing systems, a similar layer-by-layer processing technique is used (Qiao et al, 2018;Krishnan et al, 2021). Thus, the memory accessing patterns could also be a side-channel vulnerability that adversaries can exploit in memristor computing systems.…”
Section: Thwarting Side-channel Attacksmentioning
confidence: 99%
“…Thus, RRAM and SRAM-based IMC accelerators have been proposed for DNNs in the literature [19,25]. However, IMC increases on-chip data volume, which increases latency and energy due to on-chip communication [26][27][28][29]. The high density and complexity of GCNs make the on-chip communication for IMC-based accelerators even more critical.…”
Section: Related Workmentioning
confidence: 99%