2020 IEEE International Parallel and Distributed Processing Symposium (IPDPS) 2020
DOI: 10.1109/ipdps47924.2020.00042
|View full text |Cite
|
Sign up to set email alerts
|

XSP: Across-Stack Profiling and Analysis of Machine Learning Models on GPUs

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
17
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 25 publications
(17 citation statements)
references
References 2 publications
0
17
0
Order By: Relevance
“…Recent researches can be largely categorized into two kinds. The first deeply studies software and hardware issues that may affect operator execution time, and consider these issues to do estimation [7,17]. While Talos [26] and DUET [32] mainly concentrate on hardware speedup issues.…”
Section: Related Workmentioning
confidence: 99%
“…Recent researches can be largely categorized into two kinds. The first deeply studies software and hardware issues that may affect operator execution time, and consider these issues to do estimation [7,17]. While Talos [26] and DUET [32] mainly concentrate on hardware speedup issues.…”
Section: Related Workmentioning
confidence: 99%
“…Researchers and engineers often have to implement this functionality for each new project from scratch, which can become very complicated particularly when targeting new hardware, embedded devices, TinyML and IoT. -Machine learning benchmarking initiatives such as MLPerf [16], MLModelScope [17] and Deep500 [18] attempt to standardize machine learning model benchmarking and make it more reproducible. However, production deployment, integration with complex systems and adaptation to continuously changing user environments, platforms, tools and data are currently out of their scope.…”
Section: Collective Knowledge Conceptmentioning
confidence: 99%
“…They are very useful for data scientists but do not yet provide a universal mechanism to automatically build and run algorithms across different platforms, environments, libraries, tools, models and datasets. Researchers and engineers often have to implement this functionality for each new project from scratch, which can become very complicated particularly when targeting new hardware, embedded devices, TinyML and IoT.Machine learning benchmarking initiatives such as MLPerf [16], MLModelScope [17] and Deep500 [18] attempt to standardize machine learning model benchmarking and make it more reproducible. However, production deployment, integration with complex systems and adaptation to continuously changing user environments, platforms, tools and data are currently out of their scope.Package managers such as Spack [19] and EasyBuild [20] are very useful for rebuilding and fixing the whole software environment.…”
Section: Collective Knowledge Conceptmentioning
confidence: 99%
See 1 more Smart Citation
“…MLModelScope supports different combinations of DL models, frameworks, and hardware devices, allows scalable evaluation, and reports informative benchmarking results. On GPU devices, MLModelScope further provides an automated analysis tool, called XSP [99], to build an integrated view of various performance-related metrics of workloads across the entire stack. Another recent benchmarking platform is called ML-Commons [100], which provides benchmarking, datasets, and practical innovative ML models.…”
Section: Scalable Benchmarking Of Models Sw and Hwmentioning
confidence: 99%