2018
DOI: 10.48550/arxiv.1812.09926
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SNAS: Stochastic Neural Architecture Search

Abstract: We propose Stochastic Neural Architecture Search (SNAS), an economical endto-end solution to Neural Architecture Search (NAS) that trains neural operation parameters and architecture distribution parameters in same round of backpropagation, while maintaining the completeness and differentiability of the NAS pipeline. In this work, NAS is reformulated as an optimization problem on parameters of a joint distribution for the search space in a cell. To leverage the gradient information in generic differentiable lo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
202
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 199 publications
(203 citation statements)
references
References 21 publications
1
202
0
Order By: Relevance
“…However, such methods require training the searched architecture from scratch for each search step, which is extremely computationally expensive. To address this, weight-sharing approaches have been proposed [4,7,8,10,29,55,66,80,86,93,103]. They train the supernet once which includes all architecture candidates.…”
Section: Neural Architecture Searchmentioning
confidence: 99%
See 3 more Smart Citations
“…However, such methods require training the searched architecture from scratch for each search step, which is extremely computationally expensive. To address this, weight-sharing approaches have been proposed [4,7,8,10,29,55,66,80,86,93,103]. They train the supernet once which includes all architecture candidates.…”
Section: Neural Architecture Searchmentioning
confidence: 99%
“…Compared to standard ANNs, SNNs require significantly higher computational cost for training due to multiple feedforward steps [53]. This makes it difficult to search for an optimal SNN architecture with NAS techniques that train the architecture candidate multiple times [2,78,[107][108][109] or train a complex supernet [8,29,55,86]. To minimize the training budget, our work is motivated by the previous works [12,58,88] which demonstrate that the optimal architecture can be founded without any training process.…”
Section: Nas Without Trainingmentioning
confidence: 99%
See 2 more Smart Citations
“…Compared with traditional manually designed models, NAS requires less labor work and achieves better performance. Existing NAS algorithms can be categorized into three groups: 1) Reinforcement learning-based approaches [28,44,45], 2) Evolution-based approaches [22,29], and 3) Gradient-based approaches [3,6,36]. For reinforcement learning approaches, candidate architectures are sampled from search space based on reinforcement learning algorithms.…”
Section: Neural Architecture Searchmentioning
confidence: 99%