2016
DOI: 10.1145/3007787.3001155
|View full text |Cite
|
Sign up to set email alerts
|

Energy efficient architecture for graph analytics accelerators

Abstract: Specialized hardware accelerators can significantly improve the performance and power efficiency of compute systems. In this paper, we focus on hardware accelerators for graph analytics applications and propose a configurable architecture template that is specifically optimized for iterative vertex-centric graph applications with irregular access patterns and asymmetric convergence. The proposed architecture addresses the limitations of the existing multi-core CPU and GPU architectures for these types of appli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
67
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 63 publications
(67 citation statements)
references
References 31 publications
(38 reference statements)
0
67
0
Order By: Relevance
“…Google's Tensor Processing Unit described above is an example that is targeted for neural network applications. Other workloads of interest that may justify ASIC accelerators include cryptography [26], compression [27], machine learning [28], database [29], and large-scale graph processing [3,30,31].…”
Section: Google's Tensor Processing Unitmentioning
confidence: 99%
See 1 more Smart Citation
“…Google's Tensor Processing Unit described above is an example that is targeted for neural network applications. Other workloads of interest that may justify ASIC accelerators include cryptography [26], compression [27], machine learning [28], database [29], and large-scale graph processing [3,30,31].…”
Section: Google's Tensor Processing Unitmentioning
confidence: 99%
“…This leads to high power consumption by 10+ superscalar cores while not doing useful work. It was shown that custom architectures that target such communication patterns have the potential to improve power efficiency by a factor of 50x or more compared to the general-purpose CPUs [3]. …”
mentioning
confidence: 99%
“…Instead, an accelerator is expected to generate many concurrent DRAM requests to be able to hide long (typically hundreds of cycles) latencies and fully utilize the available DRAM bandwidth. It has been shown that hardware accelerators can operate at power levels that are much lower than the state-ofthe-art multi-core CPUs [26].…”
Section: Introductionmentioning
confidence: 99%
“…Because of this reason, the application-specific accelerators designed using the HLS methodology of this section will be used as a baseline in our experiments (Section VI) IV. PROPOSED ARCHITECTURE The preliminary version of this paper has proposed several microarchitectural features to achieve both high throughput and high work efficiency for asynchronous execution of graph appli- cations [26]. The basic idea is to allow processing tens/hundreds of vertices/edges to be able to hide long access latencies to main memory.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation