2017 IEEE 23rd International Conference on Embedded and Real-Time Computing Systems and Applications (RTCSA) 2017
DOI: 10.1109/rtcsa.2017.8046337
|View full text |Cite
|
Sign up to set email alerts
|

FitCNN: A cloud-assisted lightweight convolutional neural network framework for mobile devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…In this section, we look at how to tailor deep learning to mobile networking applications from three perspectives, namely, mobile devices and systems, distributed data centers, and changing mobile network environments. [513] Filter size shrinking, reducing input channels and late downsampling CNN Howard et al [514] Depth-wise separable convolution CNN Zhang et al [515] Point-wise group convolution and channel shuffle CNN Zhang et al [516] Tucker decomposition AE Cao et al [517] Data parallelization by RenderScript RNN Chen et al [518] Space exploration for data reusability and kernel redundancy removal CNN Rallapalli et al [519] Memory optimizations CNN Lane et al [520] Runtime layer compression and deep architecture decomposition MLP, CNN Huynh et al [521] Caching, Tucker decomposition and computation offloading CNN Wu et al [522] Parameters quantization CNN Bhattacharya and Lane [523] Sparsification of fully-connected layers and separation of convolutional kernels MLP, CNN Georgiev et al [97] Representation sharing MLP Cho and Brand [524] Convolution operation optimization CNN Guo and Potkonjak [525] Filters and classes pruning CNN Li et al [526] Cloud assistance and incremental learning CNN Zen et al [527] Weight quantization LSTM Falcao et al [528] Parallelization and memory sharing Stacked AE Fang et al [529] Model pruning and recovery scheme CNN Xu et al [530] Reusable region lookup and reusable region propagation scheme CNN…”
Section: Tailoring Deep Learning To Mobile Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…In this section, we look at how to tailor deep learning to mobile networking applications from three perspectives, namely, mobile devices and systems, distributed data centers, and changing mobile network environments. [513] Filter size shrinking, reducing input channels and late downsampling CNN Howard et al [514] Depth-wise separable convolution CNN Zhang et al [515] Point-wise group convolution and channel shuffle CNN Zhang et al [516] Tucker decomposition AE Cao et al [517] Data parallelization by RenderScript RNN Chen et al [518] Space exploration for data reusability and kernel redundancy removal CNN Rallapalli et al [519] Memory optimizations CNN Lane et al [520] Runtime layer compression and deep architecture decomposition MLP, CNN Huynh et al [521] Caching, Tucker decomposition and computation offloading CNN Wu et al [522] Parameters quantization CNN Bhattacharya and Lane [523] Sparsification of fully-connected layers and separation of convolutional kernels MLP, CNN Georgiev et al [97] Representation sharing MLP Cho and Brand [524] Convolution operation optimization CNN Guo and Potkonjak [525] Filters and classes pruning CNN Li et al [526] Cloud assistance and incremental learning CNN Zen et al [527] Weight quantization LSTM Falcao et al [528] Parallelization and memory sharing Stacked AE Fang et al [529] Model pruning and recovery scheme CNN Xu et al [530] Reusable region lookup and reusable region propagation scheme CNN…”
Section: Tailoring Deep Learning To Mobile Networkmentioning
confidence: 99%
“…Beyond these works, researchers also successfully adapt deep learning architectures through other designs and sophisticated optimizations, such as parameters quantization [522], [527], sparsification and separation [523], representation and memory sharing [97], [528], convolution operation optimization [524], pruning [525], cloud assistance [526] and compiler optimization [532]. These techniques will be of great significance when embedding deep neural networks into mobile systems.…”
Section: A Tailoring Deep Learning To Mobile Devices and Systemsmentioning
confidence: 99%
“…We call such a node, the edge server. Instances of such an architecture recently emerged [6]- [8] to support a variety of real-time applications [9]- [11]. The recent NVIDIA AGX platform line-up is one example of today's GPU-enabled nodes designed to be the edge server in such an architecture (NVIDIA AGX is specifically marketed as the "brain" node supporting autonomous driving [12]).…”
Section: Introductionmentioning
confidence: 99%