Proceedings of the Genetic and Evolutionary Computation Conference 2018
DOI: 10.1145/3205455.3205489
|View full text |Cite
|
Sign up to set email alerts
|

Evolutionary architecture search for deep multitask networks

Abstract: Multitask learning, i.e. learning several tasks at once with the same neural network, can improve performance in each of the tasks. Designing deep neural network architectures for multitask learning is a challenge: There are many ways to tie the tasks together, and the design choices matter. The size and complexity of this problem exceeds human design ability, making it a compelling domain for evolutionary optimization. Using the existing state of the art soft ordering architecture as the starting point, metho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
66
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 88 publications
(66 citation statements)
references
References 30 publications
0
66
0
Order By: Relevance
“…In addition to the three goals of evolutionary AutoML demonstrated in this paper, a fourth one is to take advantage of multiple related datasets. As shown in prior work [30], even when there is little data to train a DNN in a particular task, other tasks in a multitask setting can help achieve good performance. Evolutionary AutoML thus forms a framework for utilizing DNNs in domains that otherwise would be impractical due to lack of data.…”
Section: Discussionmentioning
confidence: 95%
“…In addition to the three goals of evolutionary AutoML demonstrated in this paper, a fourth one is to take advantage of multiple related datasets. As shown in prior work [30], even when there is little data to train a DNN in a particular task, other tasks in a multitask setting can help achieve good performance. Evolutionary AutoML thus forms a framework for utilizing DNNs in domains that otherwise would be impractical due to lack of data.…”
Section: Discussionmentioning
confidence: 95%
“…SFGs probabilistically defines the grouping of kernels and thus the connectivity of features in a CNNs. We use variational inference to approximate the distribution (ii) (i) (iii) (iv) Our method can be considered as a probabilistic form of multi-task architecture learning [34], as the learned posterior embodies the optimal MTL architecture given the data.…”
Section: Discussionmentioning
confidence: 99%
“…A recent step in this direction is made by Liu et al [19] who propose an adaptive MTL model that structurally groups tasks together. Evolutionary algorithms have also been shown to capture task relatedness and create sharing structures [16]. A less architectural solution is proposed by Yang et al [40] who use a factorized space representation to initialize and learn intertask sharing structures at each layer in an MTL model.…”
Section: Related Workmentioning
confidence: 99%
“…This search duration grows proportionally with the number of tasks and parameters present in the model's structure. Previous works in both MTL and STL rely on evolutionary algorithms [16] or factorization techniques [40] to discover their optimal way of learning, however this takes time and prolongs the training process. In our work, inspired by the efficiency of Random Search [3] we enforce a structured random solution to this problem by regulating the per-task data-flow in our models.…”
Section: Introductionmentioning
confidence: 99%