2021
DOI: 10.1109/tpds.2020.3047974
|View full text |Cite
|
Sign up to set email alerts
|

The Case for Strong Scaling in Deep Learning: Training Large 3D CNNs with Hybrid Parallelism

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 28 publications
(18 citation statements)
references
References 26 publications
0
16
0
Order By: Relevance
“…Especially, considering 3D reconstruction often requires comparatively long computation times. As it stands, memory and computation requirements for training 3D neural networks can be orders of magnitude larger than their 2D counterparts 57 . As such, challenges regarding network size, as well as stability during training, are major hurdles for successful implementation.…”
Section: Compressed Sensingmentioning
confidence: 99%
See 1 more Smart Citation
“…Especially, considering 3D reconstruction often requires comparatively long computation times. As it stands, memory and computation requirements for training 3D neural networks can be orders of magnitude larger than their 2D counterparts 57 . As such, challenges regarding network size, as well as stability during training, are major hurdles for successful implementation.…”
Section: Compressed Sensingmentioning
confidence: 99%
“…As it stands, memory and computation requirements for training 3D neural networks can be orders of magnitude larger than their 2D counterparts. 57 As such, challenges regarding network size, as well as stability during training, are major hurdles for successful implementation. With these constraints in mind, Pham et al 52 adapt a 2D neural network (Super-Resolution Convolutional Neural Network (SRCNN) 58 ) into a 3D equivalent (SRCNN3D) to perform 3D image reconstruction.…”
Section: Compressed Sensingmentioning
confidence: 99%
“…Livermore Big Artificial Neural Network can spatially partition the training over many graphics processing unit (GPU)–accelerated HPC nodes, enabling the traditional robust scaling that other HPC applications enjoy, that is, accelerated time to solution without a compromise in the quality of the learned model (Van Essen et al, 2015). For the CosmoFlow problem, LBANN is able to achieve an order-of-magnitude improvement in prediction quality using the full 3D data sets in training while significantly reducing training time by exploiting a much larger-scale system (Oyama et al, 2020). The GAN-based surrogate models should be able to take advantage of LBANN to an even greater degree.
Figure 3.Logarithmic histograms of pixel intensities from the recurrent neural networks–generated and validation cosmology data sets show an excellent match.
…”
Section: Surrogate Modelsmentioning
confidence: 99%
“…Data parallelism has been also used for COVID-19 diagnosis based on CT scans [14], text and feature extraction based diagnosis using CNN models [15], [16]. Some studies combine some of the abovementioned techniques, called hybrid parallelism, to handle 3D images and models [17], [18]. In [19] a scalable toolkit for medical image segmentation is presented, but is privative and only two models are provided.…”
Section: State Of the Artmentioning
confidence: 99%