2019 IEEE International Symposium on Multimedia (ISM) 2019
DOI: 10.1109/ism46123.2019.00042
|View full text |Cite
|
Sign up to set email alerts
|

Resource-Efficient Object Detection by Sharing Backbone CNNs

Abstract: The detection of objects in image and video has made huge progress in recent years due to the use of deep convolutional neural networks (DNNs), with some network architectures becoming de-facto standards. This paper addresses the problem of sharing a backbone CNN for different tasks, for example, to enable detection of additional classes when an already trained network is available. When using multiple such neural networks, sharing a backbone can save inference time and memory consumption. We study sharing a c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…However, when a lightweight model is derived from a compressed version of a complex model, specific training techniques designed for such compressed models, e.g., quantizationaware training [157], can be incorporated in the retraining phase to optimize continuous learning. Moreover, multiple models can share the same backbone to reduce memory consumption [9]. In our case, spatial aggregation in aggregated model training allows multiple video streams to have the same backbone parameters, so that only one copy of model backbone can be maintained in the GPU memory on each edge server, and memory consumption can be further reduced.…”
Section: Discussionmentioning
confidence: 99%
“…However, when a lightweight model is derived from a compressed version of a complex model, specific training techniques designed for such compressed models, e.g., quantizationaware training [157], can be incorporated in the retraining phase to optimize continuous learning. Moreover, multiple models can share the same backbone to reduce memory consumption [9]. In our case, spatial aggregation in aggregated model training allows multiple video streams to have the same backbone parameters, so that only one copy of model backbone can be maintained in the GPU memory on each edge server, and memory consumption can be further reduced.…”
Section: Discussionmentioning
confidence: 99%