2022
DOI: 10.36227/techrxiv.21547503
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An Efficient CNN Inference Accelerator Based on Intra- and Inter-Channel Feature Map Compression

Abstract: <p>Deep convolutional neural networks (CNNs) generate intensive inter-layer data during inference, which results in substantial on-chip memory size and off-chip bandwidth. To solve the memory constraint, this paper proposes an accelerator adopted with a compression technique that can reduce the inter-layer data by removing both intra- and inter-channel redundant information. Principal component analysis (PCA) is utilized in the compression process to concentrate inter-channel information. The spatial dif… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 41 publications
(48 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?