2021
DOI: 10.48550/arxiv.2111.15106
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MAPLE: Microprocessor A Priori for Latency Estimation

Abstract: Modern deep neural networks must demonstrate state-of-the-art accuracy while exhibiting low latency and energy consumption. As such, neural architecture search (NAS) algorithms take these two constraints into account when generating a new architecture. However, efficiency metrics such as latency are typically hardware dependent requiring the NAS algorithm to either measure or predict the architecture latency. Measuring the latency of every evaluated architecture adds a significant amount of time to the NAS pro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(11 citation statements)
references
References 32 publications
(59 reference statements)
0
11
0
Order By: Relevance
“…In this section we evaluate and compare MAPLE-X with MAPLE [1] and HELP [4]. We adopt a one-device-leave-out approach and form a training pool of five devices listed in Section 3.1 and use the sixth device for testing.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…In this section we evaluate and compare MAPLE-X with MAPLE [1] and HELP [4]. We adopt a one-device-leave-out approach and form a training pool of five devices listed in Section 3.1 and use the sixth device for testing.…”
Section: Resultsmentioning
confidence: 99%
“…To evaluate our methodology, we use the same dataset as MAPLE which we refer to as NASBench-X [1]. NASBench-X is based on NASBench-201, a cell-based convolutional neural network NAS dataset and includes a total of 15,625 architectures.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…As an alternative to direct measurements, existing approaches for evaluating the efficiency of a neural architecture can be categorized as those using: (1) Proxy metrics [50,65] (e.g., FLOPs), which are usually platform-independent and cannot accurately reflect the actual performance due to the diversity of platforms [39,51]. (2) Lookup tables [7,11,56], which are collected for pre-defined building blocks in the search space, but cannot cover every possible configuration in a potentially huge search space and require comprehensive measurements on each platform. (3) Prediction models [2,8,12], which broadly rely on machine learning techniques (e.g., MLPs) and have the potential to predict the performance of any configuration in the search space.…”
Section: Introductionmentioning
confidence: 99%
“…(2) Lookup tables [7,11,56], which are collected for pre-defined building blocks in the search space, but cannot cover every possible configuration in a potentially huge search space and require comprehensive measurements on each platform. (3) Prediction models [2,8,12], which broadly rely on machine learning techniques (e.g., MLPs) and have the potential to predict the performance of any configuration in the search space. However, it is difficult to build accurate prediction models for efficiency metrics on mobile devices due to the following challenges.…”
Section: Introductionmentioning
confidence: 99%