2021
DOI: 10.48550/arxiv.2102.11535
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neural Architecture Search on ImageNet in Four GPU Hours: A Theoretically Inspired Perspective

Abstract: Neural Architecture Search (NAS) has been explosively studied to automate the discovery of top-performer neural networks. Current works require heavy training of supernet or intensive architecture evaluations, thus suffering from heavy resource consumption and often incurring search bias due to truncated training or approximations. Can we select the best neural architectures without involving any training and eliminate a drastic portion of the search cost? We provide an affirmative answer, by proposing a novel… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(21 citation statements)
references
References 52 publications
0
18
0
Order By: Relevance
“…In very recent works, the key focus has been the efficiency of the NAS technique [1,91,92,94,105] owing to the growing size of dataset and architecture. Interestingly, a line of work suggests the concept of NAS without training where the networks do not require training during the search stage [12,58,88]. This can significantly reduce the computational cost for searching optimal architecture.…”
Section: Neural Architecture Searchmentioning
confidence: 99%
See 4 more Smart Citations
“…In very recent works, the key focus has been the efficiency of the NAS technique [1,91,92,94,105] owing to the growing size of dataset and architecture. Interestingly, a line of work suggests the concept of NAS without training where the networks do not require training during the search stage [12,58,88]. This can significantly reduce the computational cost for searching optimal architecture.…”
Section: Neural Architecture Searchmentioning
confidence: 99%
“…This makes it difficult to search for an optimal SNN architecture with NAS techniques that train the architecture candidate multiple times [2,78,[107][108][109] or train a complex supernet [8,29,55,86]. To minimize the training budget, our work is motivated by the previous works [12,58,88] which demonstrate that the optimal architecture can be founded without any training process. Specifically, Mellor et al [58] provide the interesting observation that the architecture having distinctive representations across different data samples is likely to achieve higher posttraining performance.…”
Section: Nas Without Trainingmentioning
confidence: 99%
See 3 more Smart Citations