2019
DOI: 10.48550/arxiv.1901.10183
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Modular Benchmarking Infrastructure for High-Performance and Reproducible Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
1
1

Relationship

5
2

Authors

Journals

citations
Cited by 9 publications
(11 citation statements)
references
References 0 publications
0
11
0
Order By: Relevance
“…An important direction of future work is also considering more realistic large-scale distributed workloads (e.g., using traces), such as different Remote Direct Memory Access based applications [12,15,32,56], deep learning training and inference [5,6], communication-intense linear algebra kernels [46], or irregular processing [16][17][18]58].…”
Section: Discussion and Takeawaymentioning
confidence: 99%
See 1 more Smart Citation
“…An important direction of future work is also considering more realistic large-scale distributed workloads (e.g., using traces), such as different Remote Direct Memory Access based applications [12,15,32,56], deep learning training and inference [5,6], communication-intense linear algebra kernels [46], or irregular processing [16][17][18]58].…”
Section: Discussion and Takeawaymentioning
confidence: 99%
“…4 Inter-router cables + server links. 5 The average over the flow size distribution, excluding retransmissions.…”
Section: Memorymentioning
confidence: 99%
“…The complexity and diversity of machine learning frameworks, available hardware systems, evaluation techniques, suitable metrics for quantification, and the limited availability of appropriate scientific datasets make this a challenging endeavour. Early initiatives on this front include MLPerf [59], AI Benchmark Suites from BenchCouncil [60], CORAL-2 [61], and Deep500 [62,63].…”
Section: Introductionmentioning
confidence: 99%
“…• Finally, the Deep500 effort is predominantly focused on techniques for reliably reporting performance of deep learning applications using metrics such as scalability, throughput, communication volume and time-to-solution [62,63]. This is more focused on methodology (and a corresponding framework) for quantifying and reporting deep learning performance than on any specific application.…”
Section: Introductionmentioning
confidence: 99%
“…Large graphs are a basis of many problems in machine learning, medicine, social network analysis, computational sciences, and others [15,25,106]. The growing graph sizes, reaching one trillion edges in 2015 (the Facebook social graph [48]) and 12 trillion edges in 2018 (the Sogou webgraph [101]), require unprecedented amounts of compute power, storage, and energy.…”
Section: Introductionmentioning
confidence: 99%