2014
DOI: 10.1007/978-3-319-06486-4
|View full text |Cite
|
Sign up to set email alerts
|

High-Performance Computing on the Intel® Xeon Phi™

Abstract: The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. While the advice and information in this book are believed to be true and accurate at the date of publication, neither the authors nor the editors nor the publisher can accept any legal responsibility for any errors or omissions that… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
2
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 50 publications
(30 citation statements)
references
References 0 publications
0
30
0
Order By: Relevance
“…Several libraries and parallelization methods can be employed in order to exploit all the capabilities of the Xeon Phi. 9,11,12,34 Intel's math kernel library (MKL) provides several mathematical functions that are already optimized for the Xeon Phi coprocessor. In particular, MCsquare uses the MCG59 random number generator of this library.…”
Section: F Code Implementationmentioning
confidence: 99%
See 1 more Smart Citation
“…Several libraries and parallelization methods can be employed in order to exploit all the capabilities of the Xeon Phi. 9,11,12,34 Intel's math kernel library (MKL) provides several mathematical functions that are already optimized for the Xeon Phi coprocessor. In particular, MCsquare uses the MCG59 random number generator of this library.…”
Section: F Code Implementationmentioning
confidence: 99%
“…Intermediate approaches also exist, like the recently introduced Intel Xeon Phi coprocessors. [9][10][11][12] They are available as affordable extension cards for workstations, just like GPUs. They combine the advantages of clusters, with many independent calculation units, and those of GPUs, with access to a shared memory and vectorized calculation.…”
Section: Introductionmentioning
confidence: 99%
“…For the novel algorithms, absolute tolerances for the Lanczos iterations and the REML optimization procedure were set to 5e-5 and 1e-5, respectively. Additionally, we compared our interpreted Python 3.6 code to BOLT-LMM versions 2.1 and 2.3.3 (C++ code compiled against the Intel MKL and Boost libraries) [5,6,24,25]. We ran each algorithm twenty times per condition, trimming away the two most extreme timings in each condition.…”
Section: Numerical Experimentsmentioning
confidence: 99%
“…Novel algorithms were implemented in the Python v3.6.5 computing environment [20], using NumPy v1.14.3 and SciPy v1.1.0 compiled against the Intel Math Kernel Library v2018.0.2 [25][26][27]. Optimization was performed using SciPy's implementation of Brent's method, with convergence determined via absolute tolerance of the standardized genomic variance componentĥ 2 .…”
Section: Numerical Experimentsmentioning
confidence: 99%
“…One MKL RNG interface API call can deliver an arbitrary number of random numbers. In our program, a maximum of 64 K random numbers are delivered in one call [30]. A thread generates the required number of random numbers for each task.…”
Section: / W a It F O R A L L T A S K S T O Be F I N I S H E D / / T mentioning
confidence: 99%