2018 IEEE 20th International Workshop on Multimedia Signal Processing (MMSP) 2018
DOI: 10.1109/mmsp.2018.8547134
|View full text |Cite
|
Sign up to set email alerts
|

Near-Lossless Deep Feature Compression for Collaborative Intelligence

Abstract: Collaborative intelligence is a new paradigm for efficient deployment of deep neural networks across the mobilecloud infrastructure. By dividing the network between the mobile and the cloud, it is possible to distribute the computational workload such that the overall energy and/or latency of the system is minimized. However, this necessitates sending deep feature data from the mobile to the cloud in order to perform inference. In this work, we examine the differences between the deep feature data and natural … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
60
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 63 publications
(60 citation statements)
references
References 18 publications
(36 reference statements)
0
60
0
Order By: Relevance
“…Furthermore, pushing all computations toward the cloud can lead to congestion in a scenario where a large number of mobile devices simultaneously send data to the cloud. As a compromise between the mobile-only and the cloud-only approach, recently, a body of research work has been investigating the idea of splitting a deep inference network between the mobile and cloud [6][7][8][9][10][11][12]. In this approach, which is referred to as collaborative intelligence, the computations associated with initial layers of the inference network are performed on the mobile device, and the feature tensor (activations) of the last computed layer is sent to the cloud for the remainder of computations.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Furthermore, pushing all computations toward the cloud can lead to congestion in a scenario where a large number of mobile devices simultaneously send data to the cloud. As a compromise between the mobile-only and the cloud-only approach, recently, a body of research work has been investigating the idea of splitting a deep inference network between the mobile and cloud [6][7][8][9][10][11][12]. In this approach, which is referred to as collaborative intelligence, the computations associated with initial layers of the inference network are performed on the mobile device, and the feature tensor (activations) of the last computed layer is sent to the cloud for the remainder of computations.…”
Section: Introductionmentioning
confidence: 99%
“…In research studies investigating collaborative intelligence, a given deep network is split between the mobile device and the cloud without any modification to the network architecture itself [6,[8][9][10][11][12]. In this paper, we investigate altering the underlying deep model architecture to make it collaborative intelligence friendly.…”
Section: Introductionmentioning
confidence: 99%
“…However, this approach can cause congestion problems due to the growth of the volume of data transmitted over the network and the number of devices linked to the cloud. To address this problem, recent studies in collaborative intelligence have developed optimized deployment strategies [2][3][4][5][6][7].…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, considering the impact of the volume of data on the congestion problem, it is desirable to compress and transfer a lesser volume of data to the cloud, unless the inference performance loss is large [3]. Previous studies [4,5] have explored the efficacy of compressing deep feature tensors using conventional standard codecs, in the context of object detection and image classification. Also, [6] suggests a method to first reduce the dimensionality of the deep feature tensor, then compress it using a codec.…”
Section: Introductionmentioning
confidence: 99%
“…In CI, the mobile device runs a part of the deep model between the input and some layer, generates a set of deep features, and sends them to the cloud for further processing by the remainder of the deep model, which resides in the cloud. In this context, the issues of deep feature compression [3,4,5] and transmission [6] become important. In [3], This work was supported in part by NSERC Grant RGPIN-2016-04590.…”
Section: Introductionmentioning
confidence: 99%