The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2021
DOI: 10.1109/ojcas.2021.3072884
|View full text |Cite
|
Sign up to set email alerts
|

Lightweight Compression of Intermediate Neural Network Features for Collaborative Intelligence

Abstract: In collaborative intelligence applications, part of a deep neural network (DNN) is deployed on a lightweight device such as a mobile phone or edge device, and the remaining portion of the DNN is processed where more computing resources are available, such as in the cloud. This paper presents a novel lightweight compression technique designed specifically to quantize and compress the features output by the intermediate layer of a split DNN, without requiring any retraining of the network weights. Mathematical m… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(1 citation statement)
references
References 28 publications
(49 reference statements)
0
1
0
Order By: Relevance
“…Model pruning [26] is a commonly used model optimization technique aimed at reducing the size and complexity of deep learning models to improve their storage efficiency, computational efficiency, and generalization ability. The basic idea of model pruning is to reduce model complexity by removing redundant connections, reducing the number of parameters, or decreasing the number of layers while maintaining the model's performance on training and test data.…”
Section: Model Pruningmentioning
confidence: 99%
“…Model pruning [26] is a commonly used model optimization technique aimed at reducing the size and complexity of deep learning models to improve their storage efficiency, computational efficiency, and generalization ability. The basic idea of model pruning is to reduce model complexity by removing redundant connections, reducing the number of parameters, or decreasing the number of layers while maintaining the model's performance on training and test data.…”
Section: Model Pruningmentioning
confidence: 99%