2019
DOI: 10.48550/arxiv.1904.02877
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours

Abstract: Can we automatically design a Convolutional Network (Con-vNet) with the highest image classification accuracy under the latency constraint of a mobile device? Neural architecture search (NAS) has revolutionized the design of hardware-efficient ConvNets by automating this process. However, the NAS problem remains challenging due to the combinatorially large design space, causing a significant searching time (at least 200 GPU-hours). To alleviate this complexity, we propose Single-Path NAS, a novel differentiabl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
80
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

2
7

Authors

Journals

citations
Cited by 44 publications
(80 citation statements)
references
References 21 publications
0
80
0
Order By: Relevance
“…Moreover, such a one-shot model suffers a memory explosion problem as it subsumes all architectures, it simply becomes too big to train when the search space grows. Many one-shot variants emerge and the design and training strategies of supernet can be roughly classified into three categories: training the whole supernet based on dropconnect tricks [2], jointly training the weights of choices and network parameters (in turns) [4,11,27], and training it in a single-path way [7,22].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Moreover, such a one-shot model suffers a memory explosion problem as it subsumes all architectures, it simply becomes too big to train when the search space grows. Many one-shot variants emerge and the design and training strategies of supernet can be roughly classified into three categories: training the whole supernet based on dropconnect tricks [2], jointly training the weights of choices and network parameters (in turns) [4,11,27], and training it in a single-path way [7,22].…”
Section: Related Workmentioning
confidence: 99%
“…One-shot methods like [3,2] try to ensure that models with such inherited weights can best predict the true accuracy. Moreover, in view of huge memory consumption of a super network, current one-shot methods train only one model at each optimization step [4,22,7].…”
Section: Introductionmentioning
confidence: 99%
“…Recently researchers [2,11,34] propose to apply singlepath training to reduce the bias introduced by approximation and model simplification of the supernet. Det-NAS [4] follows this idea to search for an efficient object detection architecture.…”
Section: Neural Architecture Searchmentioning
confidence: 99%
“…Instead of training many architectures independently, the second type of methods resort to training a super-network and estimate the performance of architectures with shared weights from the supernetwork [1,34,21,4,36,3,29,9]. With the easy access to performance estimation of each sub-architecture, DARTS [21] introduced a gradient-based method to search for the best architecture in an end-to-end manner.…”
Section: Related Workmentioning
confidence: 99%