2008 IEEE MTT-S International Microwave Symposium Digest 2008
DOI: 10.1109/mwsym.2008.4633126
|View full text |Cite
|
Sign up to set email alerts
|

Massively parallel two-dimensional TLM algorithm on graphics processing units

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2009
2009
2016
2016

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(12 citation statements)
references
References 13 publications
0
12
0
Order By: Relevance
“…A CUDA kernel consists of special instructions for execution on the GPU's multiprocessors; details about CUDA enabled GPU architecture and its SDK are given in [6]. One of the challenges in designing kernels is to ensure processes are synchronized across all multi-processors.…”
Section: Transmission Line Matrix Algorithmsmentioning
confidence: 99%
“…A CUDA kernel consists of special instructions for execution on the GPU's multiprocessors; details about CUDA enabled GPU architecture and its SDK are given in [6]. One of the challenges in designing kernels is to ensure processes are synchronized across all multi-processors.…”
Section: Transmission Line Matrix Algorithmsmentioning
confidence: 99%
“…The GPU based TLM engines reported in [12]- [13] have been updated to work with a newly developed Qt user interface framework. The updated engine is implemented using OpenCL in C++.…”
Section: Implementation Of Tlm On Gpumentioning
confidence: 99%
“…Asynchronous parallel executions are implemented using Qt's thread objects, QThread; these objects can be used in conjunction with OpenMP and MPI to create parallel execution on multiple CPU cores or workstations. The GPU based TLM engine uses the data mapping concepts reported in [12] and [13]; figure 1. Moreover, much work has been invested to realize data locality and coalescence on the GPU.…”
Section: Implementation Of Tlm On Gpumentioning
confidence: 99%
“…Parallelism is the future of computing [9] and the interest of the Antennas and Propagation community in topics of high performance computing and, in particular, of parallel programming on GPUs to face computationally burdened problems has been remarkable, as witnessed by [2][3][4][5][6][7] and by other electromagnetic numerical methods which have benefitted from GPU computing [10][11][12][13][14][15][16][17]. From this starting point, it is clear that the electromagnetic community can take advantage of this technological evolution to employ ever-more sophisticated numerical methods.…”
Section: Introductionmentioning
confidence: 99%