2020
DOI: 10.1007/978-3-030-63618-0_10
|View full text |Cite
|
Sign up to set email alerts
|

Can We Avoid Rounding-Error Estimation in HPC Codes and Still Get Trustworthy Results?

Abstract: Numerical validation enables one to ensure the reliability of numerical computations that rely on floating-point operations. Discrete Stochastic Arithmetic (DSA) makes it possible to validate the accuracy of floating-point computations using random rounding. However, it may bring a large performance overhead compared with the standard floatingpoint operations. In this article, we show that with perturbed data it is possible to use standard floating-point arithmetic instead of DSA for the purpose of numerical v… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 15 publications
0
4
0
Order By: Relevance
“…In the Update routine, the algorithm needs to calculate the sum of data instances in each cluster and then divide the sum by the number of instances in the cluster. Therefore, when a large number of instances are added together one by one naively, the accumulation of rounding errors that may occur finally impairs the clustering quality (see Reference 24 for more illustration of the effect of rounding errors). On the other hand, using double precision ( 64‐bits arithmetic ) can reduce the effect of rounding errors to a satisfying level of accuracy in our use case, but the computational cost is higher (see, e.g., Reference 25).…”
Section: Optimizing Parallel K‐means Algorithmmentioning
confidence: 99%
See 2 more Smart Citations
“…In the Update routine, the algorithm needs to calculate the sum of data instances in each cluster and then divide the sum by the number of instances in the cluster. Therefore, when a large number of instances are added together one by one naively, the accumulation of rounding errors that may occur finally impairs the clustering quality (see Reference 24 for more illustration of the effect of rounding errors). On the other hand, using double precision ( 64‐bits arithmetic ) can reduce the effect of rounding errors to a satisfying level of accuracy in our use case, but the computational cost is higher (see, e.g., Reference 25).…”
Section: Optimizing Parallel K‐means Algorithmmentioning
confidence: 99%
“…Then the nearest centroid for each instance can be found and recorded (lines [16][17][18]. Finally the cluster label of each instance is updated according to its nearest centroid, and the changes of label are counted into the private track of each thread (lines [22][23][24][25]. The reduction directive sums the private track of all threads (line 3).…”
Section: Computeassign Routinementioning
confidence: 99%
See 1 more Smart Citation
“…Generally, FMA achieves matrix multiplication which is widely used in Convolutional Neural Networks (CNN) [19] and basic linear algebra subprograms (BLAS) [21,22]. Hardware implementations for matrix multiplication have problems with the requirement of many hardware resources [19,20], software interferences [21], memory requirements, and numerical inefficacy [23,24]. Large-size hardware shifters and CSA are required for conventional FMA architectures.…”
Section: Introductionmentioning
confidence: 99%