2019
DOI: 10.48550/arxiv.1903.09900
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

sharpDARTS: Faster and More Accurate Differentiable Architecture Search

Abstract: Neural Architecture Search (NAS) has been a source of dramatic improvements in neural network design, with recent results meeting or exceeding the performance of hand-tuned architectures. However, our understanding of how to represent the search space for neural net architectures and how to search that space efficiently are both still in their infancy.We have performed an in-depth analysis to identify limitations in a widely used search space and a recent architecture search method, Differentiable Architecture… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(22 citation statements)
references
References 16 publications
(46 reference statements)
0
22
0
Order By: Relevance
“…Searching Phase's CO2 emissions: In order to determine the amount of CO2 emissions from CV-NAS search-ing phase, we used a Strubell et al [48] inspired methodology. We begin by collecting the 45 major CV-NAS papers in the last three years (2018-2020) [11,7,61,13,37,10,42,15,21,55,60,8,56,24,25,31,38,14,59,26,53,17,45,44,46,16,57,20,22,29,36,50,35,27,43,12,51,41,40,19,18,32,58,39,62] finding 157 models. For every model, we extract the Top-1 Accuracy, Parameters, FLOPS, GPU hours and GPU type.…”
Section: Methods and Resultsmentioning
confidence: 99%
“…Searching Phase's CO2 emissions: In order to determine the amount of CO2 emissions from CV-NAS search-ing phase, we used a Strubell et al [48] inspired methodology. We begin by collecting the 45 major CV-NAS papers in the last three years (2018-2020) [11,7,61,13,37,10,42,15,21,55,60,8,56,24,25,31,38,14,59,26,53,17,45,44,46,16,57,20,22,29,36,50,35,27,43,12,51,41,40,19,18,32,58,39,62] finding 157 models. For every model, we extract the Top-1 Accuracy, Parameters, FLOPS, GPU hours and GPU type.…”
Section: Methods and Resultsmentioning
confidence: 99%
“…We argue that a generic alternate optimization of network weights and architecture weights, as suggested in previous works, e.g. [17,25], is not suitable for the unique structure of the architecture space. Hence, we design a tailor-made optimizer for this task, inspired by PEA theory.…”
Section: Xnas: Experts Neural Architecture Searchmentioning
confidence: 86%
“…Early NAS methods adopt reinforcement learning (RL) or evolutionary strategy [38,2,3,31,30,39] to search among thousands of individually trained networks, which costs huge computation sources. Recent works focus on efficient weight-sharing methods, which falls into two categories: one-shot approaches [6,4,1,7,18,33,29] and gradient-based approaches [32,27,9,8,20,12,34,23], achieve state-of-the-art results on a series of tasks [10,17,24,35,16,28] in various search spaces. They construct a super network/graph which shares weights with all sub-network/graphs.…”
Section: Related Workmentioning
confidence: 99%
“…The search phase costs 4 GPUs for about 28 hours on NVIDIA GeForce RTX 2080ti. In the retraining phase, we adopt the training strategy as previous works [20] to train the searched architecture from scratch, without any additional module. The whole process lasts 250 epochs, using SGD optimizer with a momentum of 0.9, a weight decay of 3×10 −5 .…”
Section: Dartsmentioning
confidence: 99%
See 1 more Smart Citation