2021 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW) 2021
DOI: 10.1109/ipdpsw52791.2021.00019
|View full text |Cite
|
Sign up to set email alerts
|

Practice and Experience in using Parallel and Scalable Machine Learning with Heterogenous Modular Supercomputing Architectures

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 32 publications
0
5
0
Order By: Relevance
“…Parallel learning [60] • The objective is to accelerate the learning procedure and scale up the scheme.…”
Section: Methodsmentioning
confidence: 99%
“…Parallel learning [60] • The objective is to accelerate the learning procedure and scale up the scheme.…”
Section: Methodsmentioning
confidence: 99%
“…In the GRU model, kernel_initializer is glorot_uni f orm, and the learning rate is 0.001. Since the model training runs on the JUWELS-BOOSTER [33] and DEEP-DAM [21] machines, a distribution strategy from the TensorFlow interface to distribute the training across multiple GPU with custom training loops is applied [34]. The training has been set up to use 1 to 4 GPU on one node.…”
Section: Forecasting Model Set Up and Parallel Computingmentioning
confidence: 99%
“…However, the prediction model only relies on the velocity and location time series, and the training does not include parameters such as particle size, turbulence intensity, gravity, and strain rate. The parallel computing machines JUWELS-BOOSTER and DEEP-DAM [21] from the Jülich Supercomputer Centre are used to accelerate the GRU model training process. Hence, this manuscript is organized as follows.…”
Section: Introductionmentioning
confidence: 99%
“…The MinMaxScaler is a type of scaler that scales the minimum and maximum values to be 0 and 1, respectively [30]. Since the modeling was implemented on the DEEP-DAM module [31] parallel computing machine, we have applied a distributed strategy application programming interface from the TensorFlow platform abstraction to distribute the training across multiple custom training loops [32]. The strategy has been set up with one to four GPUs on one node.…”
Section: Lstm and Gru Model Set Upmentioning
confidence: 99%