2018
DOI: 10.1007/978-3-030-01249-6_18
|View full text |Cite
|
Sign up to set email alerts
|

NetAdapt: Platform-Aware Neural Network Adaptation for Mobile Applications

Abstract: This work proposes an algorithm, called NetAdapt, that automatically adapts a pre-trained deep neural network to a mobile platform given a resource budget. While many existing algorithms simplify networks based on the number of MACs or weights, optimizing those indirect metrics may not necessarily reduce the direct metrics, such as latency and energy consumption. To solve this problem, NetAdapt incorporates direct metrics into its adaptation algorithm. These direct metrics are evaluated using empirical measure… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

3
323
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 418 publications
(341 citation statements)
references
References 24 publications
3
323
0
Order By: Relevance
“…To reduce the computational cost of search, differentiable architecture search framework is used in [28,5,45] with gradient-based optimization. Focusing on adapting existing networks to constrained mobile platforms, [48,15,12] proposed more efficient automated network simplification algorithms.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To reduce the computational cost of search, differentiable architecture search framework is used in [28,5,45] with gradient-based optimization. Focusing on adapting existing networks to constrained mobile platforms, [48,15,12] proposed more efficient automated network simplification algorithms.…”
Section: Related Workmentioning
confidence: 99%
“…Network search has shown itself to be a very powerful tool for discovering and optimizing network architectures [53,43,5,48]. For MobileNetV3 we use platform-aware NAS to search for the global network structures by optimizing each network block.…”
Section: Network Searchmentioning
confidence: 99%
“…State-of-the-art network compression methods can achieve significant reductions in network size, in some cases by an order of magnitude, but often require specialized software or hardware support. For example, unstructured pruning requires optimized sparse matrix multiplication routines to realize practical acceleration [26], platform constraint-aware compression [2,36,37] requires hardware simulators or empirical measurements, and arbitrarybit quantization [9,17] requires specialized hardware. One of the advantages of knowledge distillation is that it is easily implemented in any off-the-shelf deep learning framework without the need for extra software or hardware.…”
Section: Related Workmentioning
confidence: 99%
“…Currently, the design of computationally efficient CNNs is moving from manual tuning [17][18][19] towards automatic algorithms [20][21][22][23][24][25][26]. The incorporation of specific platform constraints to such approaches involves modeling how the network architecture relates with the optimization target.…”
Section: Related Workmentioning
confidence: 99%