2020
DOI: 10.1109/tnnls.2019.2910073
|View full text |Cite
|
Sign up to set email alerts
|

Compact and Computationally Efficient Representation of Deep Neural Networks

Abstract: At the core of any inference procedure in deep neural networks are dot product operations, which are the component that require the highest computational resources. For instance, deep neural networks such as VGG-16 require up to 15 gigaoperations in order to perform the dot products present in a single forward pass, which results in significant energy consumption and therefore limit their use in resource-limited environments, e.g., on embedded devices or smartphones. A common approach to reduce the cost of inf… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
43
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
2

Relationship

3
5

Authors

Journals

citations
Cited by 64 publications
(43 citation statements)
references
References 40 publications
0
43
0
Order By: Relevance
“…These matrix data structures do not only offer compression gains, but also an efficient execution of the associated dot product algorithm [54]. Similarly, [14] proposed two novel matrix representation, the Compressed Entropy Row (CER) and Compressed Shared Elements Row (CSER) representations, that are provably more optimal than the CSR with regards to both, compression and execution efficiency when the networks parameters have low entropy statistics.…”
Section: B Lossless Neural Network Compressionmentioning
confidence: 99%
“…These matrix data structures do not only offer compression gains, but also an efficient execution of the associated dot product algorithm [54]. Similarly, [14] proposed two novel matrix representation, the Compressed Entropy Row (CER) and Compressed Shared Elements Row (CSER) representations, that are provably more optimal than the CSR with regards to both, compression and execution efficiency when the networks parameters have low entropy statistics.…”
Section: B Lossless Neural Network Compressionmentioning
confidence: 99%
“…The computational cost of the in-place additions is negligible. Note that model compression [23,24] and efficient representations [25] can further reduce the computational costs. Fig.…”
Section: Multi-kernel Prediction Networkmentioning
confidence: 99%
“…(a) input burst (b) KPN[14] (c) KPN L25 (d) MKPN (e) ground truth Example of denoising an image of a bear at Gain ∝ 4. The detailed fur is recovered best by MKPN.…”
mentioning
confidence: 99%
“…24 Complimentary to work on higher-precision efficient hardware implementation, as presented here, efforts on improving performance of low-precision networks have shown considerable progress recently. [25][26][27] Currently, these methods require off-chip processing during training and do not target online on-device learning in neuromorphic hardware.…”
Section: Introductionmentioning
confidence: 99%