Proceedings of the 24th ACM International Conference on Multimedia 2016
DOI: 10.1145/2964284.2973801
|View full text |Cite
|
Sign up to set email alerts
|

CNNdroid

Abstract: Many mobile applications running on smartphones and wearable devices would potentially benefit from the accuracy and scalability of deep CNN-based machine learning algorithms. However, performance and energy consumption limitations make the execution of such computationally intensive algorithms on mobile devices prohibitive. We present a GPUaccelerated library, dubbed CNNdroid [1], for execution of trained deep CNNs on Android-based mobile devices. Empirical evaluations show that CNNdroid achieves up to 60X sp… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 72 publications
(6 citation statements)
references
References 8 publications
0
5
0
Order By: Relevance
“…While machine learning frameworks implement different tools and methods to automatically optimize their configurations, many parameters still require manual tuning. For example, finding an optimal batch size to maximize 1 throughput with latency constraints requires direct measurements of the inference performance on the execution platform (see Figures 5,6,10,11,15,18). Manual placement of operations between GPU and CPU can also significantly improve the execution performance of an inference model (Section 5.1).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…While machine learning frameworks implement different tools and methods to automatically optimize their configurations, many parameters still require manual tuning. For example, finding an optimal batch size to maximize 1 throughput with latency constraints requires direct measurements of the inference performance on the execution platform (see Figures 5,6,10,11,15,18). Manual placement of operations between GPU and CPU can also significantly improve the execution performance of an inference model (Section 5.1).…”
Section: Discussionmentioning
confidence: 99%
“…Recently Qualcomm announced hardware acceleration support for TensorFlow using their latest Snapdragon SoC [3]. Some research prototypes that leverage mobile device special purpose processors (e.g., DSP, GPU) also exist [13,[15][16][17][18]. Other recent research has looked at the computational behavior of CNNs and the impact of the neural network architecture, such as number of layers, depth, etc., on it [5,11,27].…”
Section: Related Workmentioning
confidence: 99%
“…Besides, a GPU-accelerated library called CNNdroid has been introduced. This library can execute trained CNN on Android-based mobile devices (Oskouei et al, 2015).…”
Section: Deep Neural Networkmentioning
confidence: 99%
“…These GPUs have parallel processing capabilities which can be exploited to accelerate CNN computations on mobile devices. Moreover, an open source, GPU accelerated, library has recently become available on github [75]. Apart from this there is also a neural compute stick (Movidius) available on the market which shows promising results for the use of some CNN on low power devices [76].…”
Section: Existing Sbds Useful For Space Missionsmentioning
confidence: 99%