2019
DOI: 10.1109/tpds.2019.2913833
|View full text |Cite
|
Sign up to set email alerts
|

Fast Deep Neural Network Training on Distributed Systems and Cloud TPUs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 52 publications
(21 citation statements)
references
References 15 publications
0
20
0
1
Order By: Relevance
“…When the batch size increases, the hardware utilization also increases, and the number of iterations for training decreases, so the training time is accelerated [19]. However, a large batch size reduces accuracy so it should be mitigated [3].…”
Section: A Training Large-scale Dnn Modelsmentioning
confidence: 99%
“…When the batch size increases, the hardware utilization also increases, and the number of iterations for training decreases, so the training time is accelerated [19]. However, a large batch size reduces accuracy so it should be mitigated [3].…”
Section: A Training Large-scale Dnn Modelsmentioning
confidence: 99%
“…We use 1e −5 as the weight for L 2 regularization. We train with a batch size of 4096, using a dropout of 0.3 on 32 TPU (You et al, 2019) cores.…”
Section: Training Detailsmentioning
confidence: 99%
“…In July 2018, Google announced edge TPUs designed for neural networks inference and training on edge computing [39]. They give high performance under small physical and power limitation.…”
Section: Edge Tpusmentioning
confidence: 99%