2017
DOI: 10.1016/j.procs.2017.05.074
|View full text |Cite
|
Sign up to set email alerts
|

Distributed training strategies for a computer vision deep learning algorithm on a distributed GPU cluster

Abstract: Deep learning algorithms base their success on building high learning capacity models with millions of parameters that are tuned in a data-driven fashion. These models are trained by processing millions of examples, so that the development of more accurate algorithms is usually limited by the throughput of the computing devices on which they are trained. In this work, we explore how the training of a state-of-the-art neural network for computer vision can be parallelized on a distributed GPU cluster. The effec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
3
2

Relationship

3
6

Authors

Journals

citations
Cited by 37 publications
(11 citation statements)
references
References 4 publications
0
11
0
Order By: Relevance
“…[26] It is obvious that having access to GPU clusters is a must to deploy deep networks in practice. [272829] Pathology laboratories, however, are already under immense financial pressure to adopt WSI technology, and acquiring and storing gigapixel histopathological scans is a formidable challenge to the adoption of digital pathology. Asking for GPUs, as a prerequisite for training or using deep AI solutions, is consequently going to be financially limiting in the foreseeable future.…”
Section: Challengesmentioning
confidence: 99%
“…[26] It is obvious that having access to GPU clusters is a must to deploy deep networks in practice. [272829] Pathology laboratories, however, are already under immense financial pressure to adopt WSI technology, and acquiring and storing gigapixel histopathological scans is a formidable challenge to the adoption of digital pathology. Asking for GPUs, as a prerequisite for training or using deep AI solutions, is consequently going to be financially limiting in the foreseeable future.…”
Section: Challengesmentioning
confidence: 99%
“…A learning rate value proportional to the batch size, warmup learning rate behaviour, batch normalization, SGD to RMSProp optimizer transition are some of the techniques exposed in these works. A study of the distributed training methods using ResNet-50 architecture on a HPC cluster is shown in [10,11]. To know more about the algorithms used in this field we refer to [8].…”
Section: Related Workmentioning
confidence: 99%
“…2) Distributed Deep Learning: In previous works [7], we explored how distributed learning can help to speed up training for neural networks. Several work on spatial, model and data parallelism has been done during the recent years [8], [9], [10], including the implementation of these techniques into the most used deep learning frameworks, e.g.…”
Section: State Of the Artmentioning
confidence: 99%