2023
DOI: 10.1007/s10115-023-01891-w
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid CPU/GPU/APU accelerated query, insert, update and erase operations in hash tables with string keys

Abstract: Modern computer systems can use different types of hardware acceleration to achieve massive performance improvements. Some accelerators like FPGA and dedicated GPU (dGPU) need optimized data structures for the best performance and often use dedicated memory. In contrast, APUs, which are a combination of a CPU and an integrated GPU (iGPU), support shared memory and allow the iGPU to work together with the CPU on pointer-based data structures. First, we develop an approach for dGPU to accelerate queries in libcu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 22 publications
0
0
0
Order By: Relevance
“…Additionally, this architecture minimizes memory latency by consolidating all data in shared and global memory, contrasting with CPU only which incurs latency when loading/unloading data. This strategic choice enhances the algorithm's overall performance on parallel architectures as the recent article published by (Groth et al, 2023) in the knowledge and information systems journal has proved to be 40% higher throughput than the CPU-only approach when dealing with string manipulation operations. Our approach offers a considerable speedup, especially when dealing with a large data set of patterns as mentioned in the 2023 study published Application of soft computing (Baloi et al, 2023) which used GPU-accelerated pattern matching to find similarity metrics.…”
Section: Introductionmentioning
confidence: 99%
“…Additionally, this architecture minimizes memory latency by consolidating all data in shared and global memory, contrasting with CPU only which incurs latency when loading/unloading data. This strategic choice enhances the algorithm's overall performance on parallel architectures as the recent article published by (Groth et al, 2023) in the knowledge and information systems journal has proved to be 40% higher throughput than the CPU-only approach when dealing with string manipulation operations. Our approach offers a considerable speedup, especially when dealing with a large data set of patterns as mentioned in the 2023 study published Application of soft computing (Baloi et al, 2023) which used GPU-accelerated pattern matching to find similarity metrics.…”
Section: Introductionmentioning
confidence: 99%