2019
DOI: 10.1109/access.2019.2952577
|View full text |Cite
|
Sign up to set email alerts
|

Multilevel Neural Network for Reducing Expected Inference Time

Abstract: It is widely known that deep neural networks (DNNs) can perform well in many applications, and can sometimes exceed human ability. However, their cost limits their impact in a variety of realworld applications, such as IoT and mobile computing. Recently, many DNN compression and acceleration methods have been employed to overcome this problem. Most methods succeed in reducing the number of parameters and FLOPs, but only a few can speed up expected inference times because of either the overhead generated from u… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
references
References 29 publications
(47 reference statements)
0
0
0
Order By: Relevance