2015 IEEE High Performance Extreme Computing Conference (HPEC) 2015
DOI: 10.1109/hpec.2015.7322467
|View full text |Cite
|
Sign up to set email alerts
|

Accelerating K-Means clustering with parallel implementations and GPU computing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
33
0
2

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 56 publications
(35 citation statements)
references
References 9 publications
0
33
0
2
Order By: Relevance
“…Also exploring K-means on GPUs [16,17] has been done. But latest FPGA implementations of K-means date back more than fifteen years [18].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Also exploring K-means on GPUs [16,17] has been done. But latest FPGA implementations of K-means date back more than fifteen years [18].…”
Section: Related Workmentioning
confidence: 99%
“…The number of iterations used are 10, 50. The results of software sequential and parallel methods are referred from Janki's work [16].…”
Section: Awsmentioning
confidence: 99%
“…We report results against a heterogeneous node based approach running a custom implementation of parallel k-means on ten heterogeneous nodes, each node consisting of an NVIDIA Tesla K20M GPU with two Intel Xeon E5-2620 CPUs [35]. Further, we compare against two GPU based implementations running on an NVIDIA Tesla K20M GPU and an NVIDIA Tesla K20C GPU respectively [4], [26], an FPGA based approach running a custom parallel k-means implementation on Xilinx ZC706 FPGA [29], and a multi-core processor based approach running a custom implementation of parallel k-means on 8-core Intel i7-3770k processor [15]. The proposed approach running on the Sunway Taihu-Light supercomputer achieves more than 100x speedup over the high-performance heterogeneous nodes based approach, between 50x-70x speedup than those single GPU based approaches, and 31x speedup over multi-core CPU based approach on their largest solvable workload sizes.…”
Section: Comparison With Other Architecturesmentioning
confidence: 99%
“…Recently, sophisticated projects have emerged in the study of Spark applications performance, such as PREDIcT [21] and RISE-2016 [12]. PREDIct is a tool including a set of prediction techniques for different areas of data analytics, while RISE2016 is a collection of scalable performance prediction techniques for big data processing in distributed multi-core systems.…”
Section: Related Workmentioning
confidence: 99%
“…A Node data structure is initialized at line 9; • In a similar manner, the following phase (lines [12][13][14][15][16] is invoked (line 51), in order to remove the lock taken earlier on the node if the the following conditions are met: i) no more tasks need to be executed, ii) no other user has locked the node, and iii) there are no other stages to start. If all the conditions are met, the lock put by the current user on the node can be released.…”
Section: Task Precedence Modelmentioning
confidence: 99%