2019
DOI: 10.3390/make1010027
|View full text |Cite
|
Sign up to set email alerts
|

Guidelines and Benchmarks for Deployment of Deep Learning Models on Smartphones as Real-Time Apps

Abstract: Deep learning solutions are being increasingly used in mobile applications. Although there are many open-source software tools for the development of deep learning solutions, there are no guidelines in one place in a unified manner for using these tools toward real-time deployment of these solutions on smartphones. From the variety of available deep learning tools, the most suited ones are used in this paper to enable real-time deployment of deep learning inference networks on smartphones. A uniform flow of im… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 37 publications
(23 citation statements)
references
References 15 publications
0
19
0
Order By: Relevance
“…As most of the smartphones are equipped with a GPU [ 37 ], Recurrent BIM-PoseNet can run in real-time on such devices, deeming it suitable for practical applications. However, there will be a lag in the camera pose estimation depending on the window length.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…As most of the smartphones are equipped with a GPU [ 37 ], Recurrent BIM-PoseNet can run in real-time on such devices, deeming it suitable for practical applications. However, there will be a lag in the camera pose estimation depending on the window length.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…ML specific options are e.g. to optimize towards the target hardware [120] regarding CPU and GPU availability, to optimize towards the target operation system (demonstrated for Android and iOS by Sehgal and Kehtarnavaz [121]) or to optimize the ML workload for a specific platform [122]. Monitoring and maintenance (see section 3.6) have to be considered in the overall architecture.…”
Section: Deploymentmentioning
confidence: 99%
“…Researches show that the number of operations in a network model can effectively estimate inference time [ 5 ]. The number of FLOPs represents how computationally expensive a model is [ 50 ]. We customize the FLOPs approach suggested by Sehgal et al [ 50 ] to calculate the computational complexity of a neural network as defined in Eq.…”
Section: Proposed Approachesmentioning
confidence: 99%
“…This particular problem was selected because the solution could be applicable to larger objects (e.g., aircrafts, trucks, ships, buildings) and generalizable across other satellite imagery datasets. We have adapted a floating-point operation (FLOP) framework [ 50 ] to measure the model’s computational complexity (G-FLOPs) and establish its correlation with time-to-predict performance. Latency reduction in real-life practical experiments was tested by adapting two leading-edge computational architectures, modern GPU and TPU.…”
Section: Introductionmentioning
confidence: 99%