The platform will undergo maintenance on Sep 14 at about 9:30 AM EST and will be unavailable for approximately 1 hour.
2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis 2010
DOI: 10.1109/sc.2010.39
|View full text |Cite
|
Sign up to set email alerts
|

Parallel Fast Gauss Transform

Abstract: We present fast adaptive parallel algorithms to compute the sum of N Gaussians at N points. Direct sequential computation of this sum would take O(N 2 ) time. The parallel time complexity estimates for our algorithms are O N np for uniform point distributions and O N np log N np + np log np for nonuniform distributions using np CPUs. We incorporate a planewave representation of the Gaussian kernel which permits "diagonal translation". We use parallel octrees and a new scheme for translating the plane-waves to … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2012
2012
2019
2019

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(13 citation statements)
references
References 26 publications
0
13
0
Order By: Relevance
“…We used ε = 0.1 and decreased the bandwidth parameter h as more cores are added to keep the number of distance computations constant per core; a similar experiment setup was used in ref. 47, though we plan to perform more thorough evaluations. The timings for the computation maintains around 60% parallel efficiency above 96 cores.…”
Section: Resultsmentioning
confidence: 99%
“…We used ε = 0.1 and decreased the bandwidth parameter h as more cores are added to keep the number of distance computations constant per core; a similar experiment setup was used in ref. 47, though we plan to perform more thorough evaluations. The timings for the computation maintains around 60% parallel efficiency above 96 cores.…”
Section: Resultsmentioning
confidence: 99%
“…Notable examples are the least-squares' approach ( [43,57]), the quantization approach and the Malliavin calculus based formulation (see [13] for a thorough comparison and improvements of these techniques). In the spirit of [17], one may also consider non-parametric regression (see [38] and [56]) combined with speeding up techniques like Kd-trees ( [32,40]) or the Fast Gauss Transform ( [61,47,50,54,51]) in the case of kernel regression.…”
Section: Outline Of the Solutionmentioning
confidence: 99%
“…We used = 0.1 and decreased the bandwidth parameter h as more cores are added to keep the number of distance computations constant per core; a similar experiment setup was used in ref. 47, though we plan to perform more thorough evaluations. The timings for the computation maintains around 60% parallel efficiency above 96 cores.…”
Section: Scalability Of Kernel Summationmentioning
confidence: 99%