2008 16th International Symposium on Field-Programmable Custom Computing Machines 2008
DOI: 10.1109/fccm.2008.19
|View full text |Cite
|
Sign up to set email alerts
|

Map-reduce as a Programming Model for Custom Computing Machines

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
36
0

Year Published

2009
2009
2018
2018

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 70 publications
(36 citation statements)
references
References 12 publications
0
36
0
Order By: Relevance
“…Memory bandwidth and the number of hard multipliers embedded in the FPGA are the constraints on mapping the inner level MapReduce pattern. Figure 4 shows results from applying our proposed piecewise GP model (1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13) to the innermost loop of MAT64. Given a memory bandwidth and the number of multipliers, the figure reveals the performance-optimal design: for example, when the memory bandwidth is 3 bytes/execution cycle and 3 multipliers are available, the performance-optimal design is (k = 3, ii = 2); when the memory bandwidth is 5 bytes/execution cycles and 5 multipliers available, the performance-optimal design is (k = 5, ii = 2), as shown in Fig.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Memory bandwidth and the number of hard multipliers embedded in the FPGA are the constraints on mapping the inner level MapReduce pattern. Figure 4 shows results from applying our proposed piecewise GP model (1)(2)(3)(4)(5)(6)(7)(8)(9)(10)(11)(12)(13) to the innermost loop of MAT64. Given a memory bandwidth and the number of multipliers, the figure reveals the performance-optimal design: for example, when the memory bandwidth is 3 bytes/execution cycle and 3 multipliers are available, the performance-optimal design is (k = 3, ii = 2); when the memory bandwidth is 5 bytes/execution cycles and 5 multipliers available, the performance-optimal design is (k = 5, ii = 2), as shown in Fig.…”
Section: Resultsmentioning
confidence: 99%
“…The Haskell functional language and Google's Sawzall are used to describe the Map and Reduce functions. Yeung et al [3] apply the MapReduce programming model to design high performance systems on FPGAs and GPUs. All these methods require designers to identify the MapReduce pattern and specify the Map and Reduce functions explicitly.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Several related research efforts focus on porting MapReduce to prominent hardware platforms for high-performance computing, including multicore processors [8], [10], [11], GPUs [6], [19] the Cell processor [7], [9] and FPGAs via direct software to hardware translation [20]. Throughout this paper, we compare our runtime system design and implementation against the design and implementation of MapReduce for the Cell proposed by de Krujif and Sankaralingam [7].…”
Section: Related Workmentioning
confidence: 99%
“…Fortunately, this situation is starting to change. Frameworks such as Map-Reduce [14] successfully exploit implicit parallelism on distributed systems and have also been extended to heterogeneous platforms such as GPU [17] and FPGA [26], but unfortunately have a restricted programming model. Other models, such as CUDA [21] and OpenCL [19], provide a restricted programming model to the users of GPU accelerators, but also expose a significant amount of hardware details.…”
Section: Introductionmentioning
confidence: 99%