2020
DOI: 10.1145/3360307
|View full text |Cite
|
Sign up to set email alerts
|

A domain-specific supercomputer for training deep neural networks

Abstract: Google's TPU supercomputers train deep neural networks 50x faster than general-purpose supercomputers running a high-performance computing benchmark.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
83
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 192 publications
(90 citation statements)
references
References 8 publications
1
83
0
Order By: Relevance
“…They show that when we use larger systolic arrays (approaching the baseline size) the average PE array utilization always decreases. This behaviour has been documented in prior studies [20], [24] and occurs due to difficulties when mapping GEMM operations with irregular shapes. Input matrices fall into this category, when they have a dimension smaller than the width of the PE array.…”
Section: Resultssupporting
confidence: 66%
“…They show that when we use larger systolic arrays (approaching the baseline size) the average PE array utilization always decreases. This behaviour has been documented in prior studies [20], [24] and occurs due to difficulties when mapping GEMM operations with irregular shapes. Input matrices fall into this category, when they have a dimension smaller than the width of the PE array.…”
Section: Resultssupporting
confidence: 66%
“…Computationally, how do we scale the training, testing and deployment of complex PIML models on large datasets efficiently, so that they perform well in a rapidly changing computational landscape [130]?…”
Section: Synthesis and Outlookmentioning
confidence: 99%
“…Jouppi et al divided the optimization process into two stages: first simplifying the optimization to TSP and then solving TSP. is is a relatively new method and achieved good results [18]. Huang et al proposed a new path merging method and path length saving value calculation formula, improved the classical saving method to solve the optimization, and improved the performance of the original algorithm [19].…”
Section: Related Workmentioning
confidence: 99%