2021
DOI: 10.48550/arxiv.2112.02958
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Automap: Towards Ergonomic Automated Parallelism for ML Models

Abstract: The rapid rise in demand for training large neural network architectures has brought into focus the need for partitioning strategies, for example by using data, model, or pipeline parallelism. Implementing these methods is increasingly supported through program primitives, but identifying efficient partitioning strategies requires expensive experimentation and expertise. We present the prototype of an automated partitioner that seamlessly integrates into existing compilers and existing user workflows. Our part… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(7 citation statements)
references
References 7 publications
0
7
0
Order By: Relevance
“…Under this setting, we report excellent parallel efficiency that remains above 99% as the number of GPUs is increased. While we have only considered data-parallelism in this study, we may be able to obtain further speed-ups by considering a combination of data-and function-parallelism techniques [54] in future studies. Figure 14 also reports the effect of batch-size of training on the resulting L 2 accuracy for the first time-window (t ∈ [0, 0.1]).…”
Section: Navier-stokes Equationmentioning
confidence: 99%
“…Under this setting, we report excellent parallel efficiency that remains above 99% as the number of GPUs is increased. While we have only considered data-parallelism in this study, we may be able to obtain further speed-ups by considering a combination of data-and function-parallelism techniques [54] in future studies. Figure 14 also reports the effect of batch-size of training on the resulting L 2 accuracy for the first time-window (t ∈ [0, 0.1]).…”
Section: Navier-stokes Equationmentioning
confidence: 99%
“…Search with PartIR actions Automap [27] applies such actions to the module, guided by a worklist consisting of the module arguments (e.g., model parameters, optimizer state and the data tensors passed to a training step function). If the compiler detects that the sharding of an argument can be propagated to another argument, then both are removed from the worklist and no longer considered for further sharding actions (See Figure 2 for an example.…”
Section: Partir Rewriting For Spmd Partitioningmentioning
confidence: 99%
“…In this work, we extend Automap [27] to support multi-axis automatic partitioning, a capability that our PartIR stack already supports through nesting tiling and reduction loops, as well as multi-axis loop propagation. However, multi-axis partitioning increases the search complexity exponentially for every axis.…”
Section: Multi-axis Searchmentioning
confidence: 99%
See 2 more Smart Citations