2020
DOI: 10.48550/arxiv.2012.01839
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Distributed Training and Optimization Of Neural Networks

Abstract: Deep learning models are yielding increasingly better performances thanks to multiple factors. To be successful, model may have large number of parameters or complex architectures and be trained on large dataset. This leads to large requirements on computing resource and turn around time, even more so when hyper-parameter optimization is done (e.g search over model architectures). While this is a challenge that goes beyond particle physics, we review the various ways to do the necessary computations in paralle… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 25 publications
(26 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?