2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00282
|View full text |Cite
|
Sign up to set email alerts
|

Deep Virtual Networks for Memory Efficient Inference of Multiple Tasks

Abstract: Deep networks consume a large amount of memory by their nature. A natural question arises can we reduce that memory requirement whilst maintaining performance. In particular, in this work we address the problem of memory efficient learning for multiple tasks. To this end, we propose a novel network architecture producing multiple networks of different configurations, termed deep virtual networks (DVNs), for different tasks. Each DVN is specialized for a single task and structured hierarchically. The hierarchic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
5
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 29 publications
(87 reference statements)
0
5
0
Order By: Relevance
“…Likewise, [22] fuses the features from the different tasks to train them jointly. The other one learns multiple tasks using a single shared architecture [11], [17], [23]- [26]. [11], [23] utilize the auxiliary tasks to rich the feature representation.…”
Section: Related Work a Multi-task Learningmentioning
confidence: 99%
“…Likewise, [22] fuses the features from the different tasks to train them jointly. The other one learns multiple tasks using a single shared architecture [11], [17], [23]- [26]. [11], [23] utilize the auxiliary tasks to rich the feature representation.…”
Section: Related Work a Multi-task Learningmentioning
confidence: 99%
“…There are several recent studies [11,23,24] that proposed a network structure in which parameters can be efficiently shared across tasks. Other approaches [15,16,22] suggest a single architecture which includes multiple internal networks (or models) so that they can assign different models to multiple tasks without increasing the parameters. However, they use a fixed model structure for each task and it requires expert efforts to assign the model to each task.…”
Section: Related Workmentioning
confidence: 99%
“…However, a single shared model for multiple tasks may cause performance degradation when associ-ated tasks are less relevant [29]. To avoid this issue, recent approaches [15,16] proposed a network architecture which can contain several sub-models to assign the them to multiple tasks. Despite their attempts for MTL, they require human efforts to construct sub-models from the network architecture and assign the model to each task.…”
Section: Introductionmentioning
confidence: 99%
“…Mallya et al studied a method for performing multiple tasks in a single deep neural network by iteratively pruning and packing the network parameters [24]. Kim et al proposed a novel architecture containing multiple networks of different configurations termed deep virtual networks with respect to different tasks and memory budgets [25]. Recently, multi-task learning with DCNNs have also been studied and applied to face attributes prediction.…”
Section: Related Workmentioning
confidence: 99%