Proceedings of the 2nd ACM SIGPLAN International Workshop on Machine Learning and Programming Languages 2018
DOI: 10.1145/3211346.3211348
|View full text |Cite
|
Sign up to set email alerts
|

Relay: a new IR for machine learning frameworks

Abstract: Machine learning powers diverse services in industry including search, translation, recommendation systems, and security. The scale and importance of these models require that they be efficient, expressive, and portable across an array of heterogeneous hardware devices. These constraints are often at odds; in order to better accommodate them we propose a new high-level intermediate representation (IR) called Relay. Relay is being designed as a purely-functional, statically-typed language with the goal of balan… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
4
2

Relationship

1
9

Authors

Journals

citations
Cited by 83 publications
(35 citation statements)
references
References 26 publications
0
31
0
Order By: Relevance
“…The deep learning [10] [2] IR) resides in the front-end, and the low-level intermediate representation (tensor IR) resides in the back-end, but is relatively traditional. TVM can better obtain the overall information of the application and complete specific optimizations (such as graph optimization) for deep learning.…”
Section: Tvm Compilation Architecture and Matrix-dsp Analysis 21 Tvmmentioning
confidence: 99%
“…The deep learning [10] [2] IR) resides in the front-end, and the low-level intermediate representation (tensor IR) resides in the back-end, but is relatively traditional. TVM can better obtain the overall information of the application and complete specific optimizations (such as graph optimization) for deep learning.…”
Section: Tvm Compilation Architecture and Matrix-dsp Analysis 21 Tvmmentioning
confidence: 99%
“…TVM [21] is an optimizing compiler architecture with a large open source community and probably the highest number of supported hardware architectures. It features integration for TensorFlow and PyTorch, but so far only supports inference workloads (their high level IR "Relay" already supports training but not the lower level implementations).…”
Section: Optimizing Compilers and Middlewarementioning
confidence: 99%
“…To explore hardware-software splits, we begin with ML workloads written in Relay, which is the intermediate representation used by the TVM compiler [1]. Relay represents a machine learning workload as a series of kernel calls, but does not make explicit the underlying hardware and software components described above.…”
Section: Overview Of Solutionmentioning
confidence: 99%