2019
DOI: 10.48550/arxiv.1906.08031
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

XNAS: Neural Architecture Search with Expert Advice

Abstract: This paper introduces a novel optimization method for differential neural architecture search, based on the theory of prediction with expert advice. Its optimization criterion is well fitted for an architecture-selection, i.e., it minimizes the regret incurred by a sub-optimal selection of operations. Unlike previous search relaxations, that require hard pruning of architectures, our method is designed to dynamically wipe out inferior architectures and enhance superior ones. It achieves an optimal worst-case r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(8 citation statements)
references
References 40 publications
0
8
0
Order By: Relevance
“…In the wake of the DARTS's open-sourcing [18], a diverse number of its variants emerge in the neural architecture search community. Some of them extend its use in higher-level architecture search spaces with performance awareness in mind [4,29], some learn a stochastic distribution instead of architectural parameters [29,30,33,7,8], and others offer remedies on discovering its lack of robustness [21,5,16,13,32].…”
Section: Introductionmentioning
confidence: 99%
“…In the wake of the DARTS's open-sourcing [18], a diverse number of its variants emerge in the neural architecture search community. Some of them extend its use in higher-level architecture search spaces with performance awareness in mind [4,29], some learn a stochastic distribution instead of architectural parameters [29,30,33,7,8], and others offer remedies on discovering its lack of robustness [21,5,16,13,32].…”
Section: Introductionmentioning
confidence: 99%
“…= f (x; α, ω), where α and ω indicate the architectural parameters and network weights, respectively. DARTS [22] and its variants [4,25,37] have relied on many manually designed rules to determine the final architecture. Examples include each edge can only preserve one operator, each inner node can preserve two of its precursors, and the architecture is shared by the same type (normal and reduction) of cells.…”
Section: Breaking the Rules: Enlarging The Search Spacementioning
confidence: 99%
“…Many recent works [2,1,3] aim to reduce the search time by training a single over-parameterized network with inherited/shared weights. For example, DARTS [27] and its variants [50,9,30,5,23,58,54,6] relaxed the architecture representation to the continuous domain to make the search differentiable. Recenly, NAS has been explored in 3D [23,43].…”
Section: Related Workmentioning
confidence: 99%
“…Although encouraging in their results, early NAS approaches were limited by their massive computational needs. Recent works like DARTS [27] and its variants [50,9,30,5,23,58,54,6] formulate architecture search as a differentiable problem, which greatly alleviates computational complexity. Further advancements are proposed in recent work like SGAS [23], where optimal architectures show better generalization between search and final tasks.…”
Section: Introductionmentioning
confidence: 99%