2016 7th International Conference on Cloud Computing and Big Data (CCBD) 2016
DOI: 10.1109/ccbd.2016.029
|View full text |Cite
|
Sign up to set email alerts
|

Benchmarking State-of-the-Art Deep Learning Software Tools

Abstract: Deep learning has been shown as a successful machine learning method for a variety of tasks, and its popularity results in numerous open-source deep learning software tools. Training a deep network is usually a very time-consuming process. To address the computational challenge in deep learning, many tools exploit hardware features such as multi-core CPUs and many-core GPUs to shorten the training time. However, different tools exhibit different features and running performance when training different types of… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
153
0
3

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 254 publications
(166 citation statements)
references
References 23 publications
(40 reference statements)
3
153
0
3
Order By: Relevance
“…For readers with interest in deep learning software tools, in 29 we can find an interesting experimental analysis on CPU and GPU platforms for some well known GPU-accelerated tools, including: Caffe (developed by the Berkeley Vision and Learning Center), CNTK (developed by Microsoft Research), Tensorflow (developed by Google) and Torch.…”
Section: Deep Learning Softwarementioning
confidence: 99%
“…For readers with interest in deep learning software tools, in 29 we can find an interesting experimental analysis on CPU and GPU platforms for some well known GPU-accelerated tools, including: Caffe (developed by the Berkeley Vision and Learning Center), CNTK (developed by Microsoft Research), Tensorflow (developed by Google) and Torch.…”
Section: Deep Learning Softwarementioning
confidence: 99%
“…The selection of the Caffe framework and the training scheme follows the common acknowledgment in the deep learning community [36,49]. Although there are plenty of selection of deep learning framework to use (such as TensorFlow, Torch), the accuracy-wise performance has only a very limited variation [49].…”
Section: Methodsmentioning
confidence: 99%
“…Although there are plenty of selection of deep learning framework to use (such as TensorFlow, Torch), the accuracy-wise performance has only a very limited variation [49]. The choice of training scheme has also undergone extensive investigation [50] and we follow [36] because of the similar network architecture.…”
Section: Methodsmentioning
confidence: 99%
“…Workers frequently fetch the up-to-date model from the PS, make computation over the data they host, and then return gradient updates to the PS. Since DNN models are large (from thousands to billions parameters [9]), placing those worker tasks over edge-devices imply significant updates transfer over the Internet. The PS being in a central location (typically at a cloud provider), the question of inbound traffic is also crucial for pricing our proposal.…”
Section: Introductionmentioning
confidence: 99%