2016 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS) 2016
DOI: 10.1109/ispass.2016.7482073
|View full text |Cite
|
Sign up to set email alerts
|

Analyzing the energy-efficiency of sparse matrix multiplication on heterogeneous systems: A comparative study of GPU, Xeon Phi and FPGA

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
17
0
1

Year Published

2016
2016
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(18 citation statements)
references
References 39 publications
0
17
0
1
Order By: Relevance
“…However, their high performance comes at the cost of high power dissipation [2]. FPGAs offer opportunities for exploiting low-level fine-grained parallelism by customizing data paths to the requirements of a specific algorithm/application [3].…”
Section: Introductionmentioning
confidence: 99%
“…However, their high performance comes at the cost of high power dissipation [2]. FPGAs offer opportunities for exploiting low-level fine-grained parallelism by customizing data paths to the requirements of a specific algorithm/application [3].…”
Section: Introductionmentioning
confidence: 99%
“…We measured a vision kernel's dynamic power while excluding the static power required to power the rest of the platform. This better reflects the actual workload that is being deployed to the system since certainly for small kernels, the compute energy [4] (energy consumed for computation only) and data transfer energy are usually dominated by the static power. In the vision pipeline evaluation, we compared the performance of HW accelerators in terms of their energy delay products (EDP).…”
Section: Benchmarking Approachmentioning
confidence: 99%
“…However, their high performance comes at the cost of high power dissipation [3]. FPGAs offer opportunities for using low-level fine-grained parallelism by customizing processing/control units and data paths to the requirements of a specific algorithm or application [4].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…For the past years, Field-Programmable Gate Arrays (FPGAs) have been shown to be promising platforms to achieve computational FLOPS/Watt performance comparable to Graphics Processing Units (GPUs) while achieving higher energy efficiency (Mittal and Vetter 2014). Such energy efficiency makes FPGA promising platforms to accelerate the next generations of Neural Networks (Nurvitadhi et al 2017), Sparse Matrix Algebra (Giefers et al 2016), network applications (Nurvitadhi et al 2016), financial market applications (Schryver et al 2011), image processing (Fowers et al 2012), and data centres (Weerasinghe et al 2015). As a result, a common configuration is to have GPP acting as a host and FPGA as a hardware accelerator.…”
Section: Introductionmentioning
confidence: 99%