2020
DOI: 10.48550/arxiv.2007.10359
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

GPU coprocessors as a service for deep learning inference in high energy physics

Jeffrey Krupa,
Kelvin Lin,
Maria Acosta Flechas
et al.

Abstract: In the next decade, the demands for computing in large scientific experiments are expected to grow tremendously. During the same time period, CPU performance increases will be limited. At the CERN Large Hadron Collider (LHC), these two issues will confront one another as the collider is upgraded for high luminosity running. Alternative processors such as graphics processing units (GPUs) can resolve this confrontation provided that algorithms can be sufficiently accelerated. In many cases, algorithmic speedups … Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1

Relationship

4
1

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 49 publications
(57 reference statements)
0
7
0
Order By: Relevance
“…Continued development in this direction may allow such algorithms to be used effectively in future computing workflows [32] and the Level-1 trigger at the LHC. In future work, we plan to study detailed comparisons of the two implementations based on the same model, as well as comparing to GPU coprocessors [33]. Other optimizations of the GNN model may also be possible, such as more efficient architectures [30] and use of quantization-aware training [31,34] to reduce the necessary precision.…”
Section: Discussionmentioning
confidence: 99%
“…Continued development in this direction may allow such algorithms to be used effectively in future computing workflows [32] and the Level-1 trigger at the LHC. In future work, we plan to study detailed comparisons of the two implementations based on the same model, as well as comparing to GPU coprocessors [33]. Other optimizations of the GNN model may also be possible, such as more efficient architectures [30] and use of quantization-aware training [31,34] to reduce the necessary precision.…”
Section: Discussionmentioning
confidence: 99%
“…The feasibility of the as-a-service computing model for HEP workflows has been previously demonstrated using SONIC to interact with a GPU-based server for inference [17]. The server/client design employed within this paper is similar to previous work, allowing for a direct comparison of the performance.…”
Section: Related Workmentioning
confidence: 99%
“…Summary of the performance of FaaST servers in terms of events and inferences per second, and bandwidth. Results for performance on GPUs are taken from Ref [17]…”
mentioning
confidence: 99%

FPGAs-as-a-Service Toolkit (FaaST)

Rankin,
Krupa,
Harris
et al. 2020
Preprint
Self Cite
“…This includes GPUs and potentially even fieldprogrammable gate arrays (FPGAs) or ML-specific processors such as the GraphCore intelligence processing units (IPUs) [27] through specialized ML compilers [28][29][30]. These coprocessing accelerators can be integrated into existing CPU-based experimental software frameworks as a scalable service that grows to meet the transient demand [31][32][33]. Fake rate…”
Section: Charged Hadronsmentioning
confidence: 99%