2019
DOI: 10.48550/arxiv.1905.08170
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DARC: Differentiable ARchitecture Compression

Abstract: In many learning situations, resources at inference time are significantly more constrained than resources at training time. This paper studies a general paradigm, called Differentiable ARchitecture Compression (DARC), that combines model compression and architecture search to learn models that are resource-efficient at inference time. Given a resource-intensive base architecture, DARC utilizes the training data to learn which sub-components can be replaced by cheaper alternatives. The high-level technique can… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
2

Relationship

2
0

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…A closely related line of work is Neural Architecture Search (NAS). It aims to efficiently search the space of architectures (Pham et al, 2018;Liu et al, 2018;Singh et al, 2019). Quantization is another technique to reduce the model size.…”
Section: Related Workmentioning
confidence: 99%
“…A closely related line of work is Neural Architecture Search (NAS). It aims to efficiently search the space of architectures (Pham et al, 2018;Liu et al, 2018;Singh et al, 2019). Quantization is another technique to reduce the model size.…”
Section: Related Workmentioning
confidence: 99%
“…Most of these works fall primarily into one of the four categories: quan- tization, low rank factorization, sparse connections, and structured pruning. Besides compression methods, there have been a lot of work in architecture search for compute efficient networks [22,29]. Recently, model compression techniques has also been applied on natural language processing models [19,17].…”
Section: Related Workmentioning
confidence: 99%