2018
DOI: 10.48550/arxiv.1810.01993
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Exascale Deep Learning for Climate Analytics

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 0 publications
0
3
0
Order By: Relevance
“…For example, the training process might often be accelerated by appropriate metrics for prediction error-loss function. The network complexity should also reflect available computational resources, because optimization and training are inevitably expensive in terms of computations (e.g., Gordon Bell prize 2018 for Kurth et al, 2018). In this section, we present a conceptual architecture of the neural network along with workflows to finding optimal parameters of the network.…”
Section: Predictor Philosophymentioning
confidence: 99%
“…For example, the training process might often be accelerated by appropriate metrics for prediction error-loss function. The network complexity should also reflect available computational resources, because optimization and training are inevitably expensive in terms of computations (e.g., Gordon Bell prize 2018 for Kurth et al, 2018). In this section, we present a conceptual architecture of the neural network along with workflows to finding optimal parameters of the network.…”
Section: Predictor Philosophymentioning
confidence: 99%
“…Therefore, recent hardware provide higher 32-bit and 16-bit processing rates at the expense of double precision. For example, the tensor cores on the Nvidia Volta architecture enable OLCF's Summit to already breach the exaops barrier [115]. Intel, in turn, has announced the Knights Mill architecture which is an Intel Xeon Phi chip with improved single precision performance compared to the previous Knights Landing architecture [116] designed specifically to serve the artificial intelligence market.…”
Section: Mixed Precisionmentioning
confidence: 99%
“…At a cost of a relatively expensive computation in the training process, deep neural networks (DNNs) provide a powerful approach to explore hidden correlations in massive data, which in many cases are physically not possible with human manual review [2]. In the past decade, the large computational cost for training DNN has been mitigated by a number of advances, including high-performance computers [3], graphics processing units (GPUs) [4], tensor processing units (TPUs) [5], and fast largescale optimization schemes [6], i.e., adaptive moment estimation (Adam) [7] and adaptive gradient algorithm (AdaGrad) [8]. In many instances for modeling physical systems, physical invariants, e.g., momentum and energy conservation laws, can be built into the learning algorithms in the context of DNN and their variants [9][10][11].…”
Section: Introductionmentioning
confidence: 99%