2019
DOI: 10.48550/arxiv.1905.12799
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Reinforcement Learning and Adaptive Sampling for Optimized DNN Compilation

Abstract: Achieving faster execution with shorter compilation time can enable further diversity and innovation in neural networks. However, the current paradigm of executing neural networks either relies on hand-optimized libraries, traditional compilation heuristics, or very recently, simulated annealing and genetic algorithms. Our work takes a unique approach by formulating compiler optimizations for neural networks as a reinforcement learning problem, whose solution takes fewer steps to converge. This solution, dubbe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 33 publications
0
4
0
Order By: Relevance
“…TensorComprehensions [79] uses genetic algorithm, AutoTVM [13], [14] uses simulated annealing and boosted tree, Reagen et. al, [59] uses Bayesian optimization, RELEASE [7] uses RL, ATLAS [84] uses black box optimizations, some compiler design [12], [50] use profile-guided optimization to perform target-independent front-end compiler optimizations on DNNs or linear algebra computations. Some recent works use RL on HW/SW co-exploration to explore both DNN and its mapping over hardware [6], [32], [44], [88].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…TensorComprehensions [79] uses genetic algorithm, AutoTVM [13], [14] uses simulated annealing and boosted tree, Reagen et. al, [59] uses Bayesian optimization, RELEASE [7] uses RL, ATLAS [84] uses black box optimizations, some compiler design [12], [50] use profile-guided optimization to perform target-independent front-end compiler optimizations on DNNs or linear algebra computations. Some recent works use RL on HW/SW co-exploration to explore both DNN and its mapping over hardware [6], [32], [44], [88].…”
Section: Related Workmentioning
confidence: 99%
“…Recently RL has been demonstrated within compilers/mappers [7], [24], [46], [51] for tiling and mapping DNNs over accelerators. ConfuciuX focuses on leveraging RL for exploring the search space during accelerator design.…”
Section: Introductionmentioning
confidence: 99%
“…To the best of our knowledge, this work is one of the first that learns the design space to generalize it. Most prior works leveraging ML for accelerator DSE [4], [7], [16], [17], [27] focus on performing the search faster. Learning the design space enables constant time prediction of the optima.…”
Section: Workload Dims Design Constraintsmentioning
confidence: 99%
“…It has also been used for designing memory systems, such as prefetching [52] and memory controller [53]. Additionally, it has been applied to DNN compilation and mapping optimization [54,55,56]. In this work, we use RL for co-exploration of data and computation mapping in NMP systems.…”
Section: Reinforcement Learning (Rl)mentioning
confidence: 99%