2019
DOI: 10.48550/arxiv.1901.05884
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

EAT-NAS: Elastic Architecture Transfer for Accelerating Large-scale Neural Architecture Search

Abstract: Neural architecture search (NAS) methods have been proposed to release human experts from tedious architecture engineering. However, most current methods are constrained in small-scale search due to the issue of computational resources. Meanwhile, directly applying architectures searched on small datasets to large datasets often bears no performance guarantee. This limitation impedes the wide use of NAS on large-scale tasks. To overcome this obstacle, we propose an elastic architecture transfer mechanism for a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1

Relationship

3
2

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 15 publications
0
6
0
Order By: Relevance
“…Therefore, Google proposed the ENAS algorithm [12], which avoids inefficient initial training by forcing all submodels to share weights, thereby improving the efficiency of ANS and overcoming the shortcomings of high computational cost and time-consuming of ANS. [20], which simultaneously evolve along network structures and hyper-parameters to generate neural networks. These methods first randomly generate a population (N sets of solutions) and begin to cycle through the following steps: selection, crossover, mutation, until the final condition is met.…”
Section: Automatic Network Searchmentioning
confidence: 99%
“…Therefore, Google proposed the ENAS algorithm [12], which avoids inefficient initial training by forcing all submodels to share weights, thereby improving the efficiency of ANS and overcoming the shortcomings of high computational cost and time-consuming of ANS. [20], which simultaneously evolve along network structures and hyper-parameters to generate neural networks. These methods first randomly generate a population (N sets of solutions) and begin to cycle through the following steps: selection, crossover, mutation, until the final condition is met.…”
Section: Automatic Network Searchmentioning
confidence: 99%
“…Following this manner, EAS (Cai et al, 2018) extends the parameter remapping concept to neural architecture search. Moreover, some NAS works (Pham et al, 2018;Fang et al, 2019a;Elsken et al, 2019) apply parameters sharing on child models to accelerate the search process. Our parameter remapping paradigm extends the mapping dimension with the kernel level.…”
Section: Related Workmentioning
confidence: 99%
“…These include meta-learning based approaches [36], [37] with application to few-shot learning tasks. XferNAS [38] and EAT-NAS [39] illustrate how architectures can be transferred between similar datasets or from smaller to larger datasets. Some approaches [40], [41] proposed RL-based NAS methods that search on multiple tasks during training and transfer the learned search strategy, as opposed to searched networks, to new tasks at inference.…”
Section: Related Workmentioning
confidence: 99%