2013 IEEE International Symposium on Parallel &Amp; Distributed Processing, Workshops and PHD Forum 2013
DOI: 10.1109/ipdpsw.2013.45
|View full text |Cite
|
Sign up to set email alerts
|

Recent Advances on GPU Computing in Operations Research

Abstract: Abstract-In the last decade, Graphics Processing Units (GPUs) have gained an increasing popularity as accelerators for High Performance Computing (HPC) applications. Recent GPUs are not only powerful graphics engines but also highly threaded parallel computing processors that can achieve sustainable speedup as compared with CPUs. In this context, researchers try to exploit the capability of this architecture to solve difficult problems in many domains in science and engineering. In this article, we present rec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 25 publications
(14 citation statements)
references
References 34 publications
0
14
0
Order By: Relevance
“…It turns out that this hierarchy is highly consistent with the hierarchy of threads and different types of memory of CUDA framework. Table 3 Correspondence between the parallel hybrid GA components and the hierarchy of According to the problem description in Section 3, a target machine matrix X(k) stored on GPU global memory with n + n′ rows and g columns, is presented in (10).  if job j at stage s is assigned to a machine after the start time of the rescheduling point, element x js (k) is equal to a random integer representing the target machine handling job j at stage s. Similarly, elements y js (k) is also generated randomly from the range starting from 1 to the amount of unassigned operations.…”
Section: Fig3 Hierarchy Of Threads and Different Types Of Memory Ofmentioning
confidence: 99%
“…It turns out that this hierarchy is highly consistent with the hierarchy of threads and different types of memory of CUDA framework. Table 3 Correspondence between the parallel hybrid GA components and the hierarchy of According to the problem description in Section 3, a target machine matrix X(k) stored on GPU global memory with n + n′ rows and g columns, is presented in (10).  if job j at stage s is assigned to a machine after the start time of the rescheduling point, element x js (k) is equal to a random integer representing the target machine handling job j at stage s. Similarly, elements y js (k) is also generated randomly from the range starting from 1 to the amount of unassigned operations.…”
Section: Fig3 Hierarchy Of Threads and Different Types Of Memory Ofmentioning
confidence: 99%
“…The experimental results showed great speedups for the exterior point algorithm and quite worse for the revised simplex method. Other valuable attempts can also be found in [36][37][38] achieving very satisfactory speedups with C1060, S1070 and GTX 670 boards.…”
Section: Related Workmentioning
confidence: 99%
“…Researchers and practitioners who develop and use metaheuristics always benefit from low cost computing power [26] GPU's are powerful accelerators, require less energy than other computing devices, are widely available and are relatively cheap. These user needs and GPU characteristics and the fact that GPU computing has been identified as a very promising direction in the field of Operations Research [4], motivated us to solve the QAP using the GPU. In our work, the intensive computational tasks in the tabu search algorithm are handled by the GPU leaving the CPU just with the tasks of reading and organizing input values and collecting output values.…”
Section: Gpu Computingmentioning
confidence: 99%
“…The use of parallel computing to solve QAP's started with the development of CPU parallel implementations such as the ones in [23,8,14]. In the last years, there has been also a shift to finding solutions using GPUs [26,4,12]. We chose to parallelize the tabu search metaheuristic because, it has been reported as the most efficient approximate method for solving the QAP [22,10,24,23].…”
Section: Introductionmentioning
confidence: 99%