Proceedings of the 15th Conference of the European Chapter of The Association for Computational Linguistics: Volume 2 2017
DOI: 10.18653/v1/e17-2026
|View full text |Cite
|
Sign up to set email alerts
|

Identifying beneficial task relations for multi-task learning in deep neural networks

Abstract: Multi-task learning (MTL) in deep neural networks for NLP has recently received increasing interest due to some compelling benefits, including its potential to efficiently regularize models and to reduce the need for labeled data. While it has brought significant improvements in a number of NLP tasks, mixed results have been reported, and little is known about the conditions under which MTL leads to gains in NLP. This paper sheds light on the specific task relations that can lead to gains from MTL models over … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

8
168
1

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 179 publications
(179 citation statements)
references
References 11 publications
8
168
1
Order By: Relevance
“…Furthermore, introducing LEX (cf. Section 4) as auxiliary task was generally helpful; on the other hand, POS did not seem to help, corroborating previous findings (Alonso and Plank, 2017;Bingel and Søgaard, 2017).…”
Section: Resultssupporting
confidence: 88%
“…Furthermore, introducing LEX (cf. Section 4) as auxiliary task was generally helpful; on the other hand, POS did not seem to help, corroborating previous findings (Alonso and Plank, 2017;Bingel and Søgaard, 2017).…”
Section: Resultssupporting
confidence: 88%
“…A common approach involves training the MTL model on different task specific corpus by randomly switching between the different tasks and updating both the task-specific and shared parameters based on its corpus. [15], [17], [19] employed this training strategy. A joint end-to-end model training strategy is mostly suitable for cases where the alternative tasks are treated as auxiliary objectives on the same dataset.…”
Section: Related Workmentioning
confidence: 99%
“…This can be beneficial in a number of scenarios. Previous work has shown benefits, e.g., in cases where one has tasks which are closely related to one another (Bjerva, 2017a,b), when one task can help another escape a local minimum (Bingel and Søgaard, 2017), and when one has access to some unsupervised signal which can be beneficial to the task at hand (Rei, 2017). A common approach to MTL is the application of hard parameter sharing, in which some set of parameters in a model is shared between several tasks.…”
Section: Related Work 21 Multitask Learningmentioning
confidence: 99%