2011
DOI: 10.1007/978-3-642-23397-5_42
|View full text |Cite
|
Sign up to set email alerts
|

Lessons Learned from Exploring the Backtracking Paradigm on the GPU

Abstract: Abstract. We explore the backtracking paradigm with properties seen as sub-optimal for GPU architectures, using as a case study the maximal clique enumeration problem, and find that the presence of these properties limit GPU performance to approximately 1.4-2.25 times a single CPU core. The GPU performance "lessons" we find critical to providing this performance include a coarse-and-fine-grain parallelization of the search space, a low-overhead load-balanced distribution of work, global memory latency hiding t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
23
1

Year Published

2013
2013
2020
2020

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 40 publications
(24 citation statements)
references
References 18 publications
0
23
1
Order By: Relevance
“…The described GPU backtracking strategy performs well in regular scenarios, but it faces a decrease of performance in more irregular ones, being outperformed even by the serial CPU implementation in some situations . The main reason for this decrease of performance is that GPUs suffer from load imbalance and diverging instruction flow.…”
Section: Background and Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…The described GPU backtracking strategy performs well in regular scenarios, but it faces a decrease of performance in more irregular ones, being outperformed even by the serial CPU implementation in some situations . The main reason for this decrease of performance is that GPUs suffer from load imbalance and diverging instruction flow.…”
Section: Background and Related Workmentioning
confidence: 99%
“…The main reason for this decrease of performance is that GPUs suffer from load imbalance and diverging instruction flow. Thus, to achieve a proper utilization of the multiprocessors, this parallel backtracking strategy must launch a huge amount of GPU threads …”
Section: Background and Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…First, GPU operations are based on warps (which are groups of threads to be executed in single-instructionmultiple-data fashion), and different execution paths generated by backtracking algorithms may cause a so-called warp divergence problem. Second, GPU implementations for coalesced memory accesses are no longer straightforward due to irregular access patterns [19].…”
Section: Introductionmentioning
confidence: 99%
“…Because of their massive data processing capability and their remarkable cost efficiency, GPUs are an attractive choice for providing the computing power needed to solve larger problem instances.The efficient implementation of the B&B algorithm on GPUs is a challenging task because the GPU programming model is at odds with the algorithm's highly irregular nature [1]. In this paper, we present a multi-GPU B&B algorithm for solving large permutation-based combinatorial optimization problems.…”
mentioning
confidence: 99%