2020
DOI: 10.1109/jproc.2020.2976475
|View full text |Cite
|
Sign up to set email alerts
|

Model Compression and Hardware Acceleration for Neural Networks: A Comprehensive Survey

Abstract: Vision Transformers (ViTs) have recently garnered considerable attention, emerging as a promising alternative to convolutional neural networks (CNNs) in several vision-related applications. However, their large model sizes and high computational and memory demands hinder deployment, especially on resource-constrained devices. This underscores the necessity of algorithm-hardware co-design specific to ViTs, aiming to optimize their performance by tailoring both the algorithmic structure and the underlying hardwa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
280
0
3

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 608 publications
(315 citation statements)
references
References 200 publications
0
280
0
3
Order By: Relevance
“…As DNN networks become deeper and more complex, the required computing power and energy consumption are also increasing [7]- [9]. Since most endpoint devices are batterypowered, energy-efficient ASICs which are able to process DNN is highly required.…”
Section: Imentioning
confidence: 99%
See 1 more Smart Citation
“…As DNN networks become deeper and more complex, the required computing power and energy consumption are also increasing [7]- [9]. Since most endpoint devices are batterypowered, energy-efficient ASICs which are able to process DNN is highly required.…”
Section: Imentioning
confidence: 99%
“…Index matching is to match the coordinate of the weight and the coordinate of the input feature map (ifmap) pixel for each operation to determine whether there is a meaningful operation that produces a non-zero product term. However, the irregular distribution of non-zero data leads to a large matching overhead in parallel processing since accelerators have to match multiple coordinates in parallel [9].…”
Section: Imentioning
confidence: 99%
“…[ 106 ] The error from each layer might accumulate and cause accuracy loss or nonconvergence. [ 144 ] In addition, end‐to‐end network adaptation for array nonideal factors is an interesting line of thinking. Hybrid training takes more account of the energy and complexity of a mapping device.…”
Section: Challenges and Outlookmentioning
confidence: 99%
“…There have also been significant strides in the development of hardware accelerators for SNNs [116], [117], [118], CNNs [119], [120], [121], GNNs [122], [123] and training accelerators [124], [125], [126]. A comprehensive survey of the topic can be found in [127], [121]. We also refer the reader to recent research on attention networks [128] used in image captioning applications, transformers [129] used in natural language processing and on neural architecture search [130] to design neural network configurations that reduce the complexity of the network.…”
Section: Conclusion and Summarymentioning
confidence: 99%