Proceedings of the 59th ACM/IEEE Design Automation Conference 2022
DOI: 10.1145/3489517.3530581
|View full text |Cite
|
Sign up to set email alerts
|

Using machine learning to optimize graph execution on NUMA machines

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 14 publications
0
4
0
Order By: Relevance
“…Traditionally, HPC servers receive batches of graphs applications to execute serially (i.e., one after another), always using all the available processing resources ‐ in the same way as any other HPC application. However, graph applications, because of their irregular structure and larger dimension, are intrinsically communication/memory‐bound, which means that they tend to use the CPU less when compared to the average parallel application 1,10 . Given that, graph applications executions are even more affected by issues that tend to limit the scalability of parallel applications related to both hardware (e.g., saturation of execution units and communication bus) and software (e.g., data synchronization and concurrent accesses to shared memory) 11‐13 .…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…Traditionally, HPC servers receive batches of graphs applications to execute serially (i.e., one after another), always using all the available processing resources ‐ in the same way as any other HPC application. However, graph applications, because of their irregular structure and larger dimension, are intrinsically communication/memory‐bound, which means that they tend to use the CPU less when compared to the average parallel application 1,10 . Given that, graph applications executions are even more affected by issues that tend to limit the scalability of parallel applications related to both hardware (e.g., saturation of execution units and communication bus) and software (e.g., data synchronization and concurrent accesses to shared memory) 11‐13 .…”
Section: Introductionmentioning
confidence: 99%
“…The rising number of cores and available memory in high-performance computing (HPC) servers has enabled the analysis of a massive amount of graph-structured data extracted from sources like Google, Facebook, and Twitter. [1][2][3] Simultaneously, the unprecedented growth of such interconnected data has been pushing forward the development of efficient graph analytics methods to extract useful information from massive data sources, making enhancements in several areas, such as business, geolocation, fraud detection, and social network analysis. 4,5 Graph algorithms such as PageRank and single source shortest-paths (SSSP) make it possible to perform operations like the page ordering of the most visited content on Google or even complex computations in the Artificial Intelligence ambit.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…These approaches, however, provide little flexibility as they require knowing the running applications before the execution starts and do not allow adapting the number of spawned threads dynamically. PredG [67] uses machine learning to select the best thread and data mapping policies to run graph applications on a NUMA system. Both approaches require knowing application-level information, such as the input graphs for decision-making.…”
Section: Coresmentioning
confidence: 99%