Proceedings of the Eighth International Workshop on Natural Language Processing for Social Media 2020
DOI: 10.18653/v1/2020.socialnlp-1.8
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Task Supervised Pretraining for Neural Domain Adaptation

Abstract: Two main transfer learning approaches are used in recent work in NLP to improve neural networks performance for under-resourced domains in terms of annotated data. 1) Multitask learning consists in training the task of interest with related tasks to exploit their underlying similarities. 2) Mono-task pretraining, where target model's parameters are pretrained on large-scale labelled source domain and then fine-tuned on labelled data from the target domain (the domain of interest). In this paper, we propose a n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(7 citation statements)
references
References 37 publications
0
7
0
Order By: Relevance
“… 3 Combining and then jointly training a small dataset of interest with a larger auxiliary dataset often helps the prediction accuracy. 22 , 23 For the auxiliary dataset, we downloaded the publicly available SIIM-ISIC Melanoma Classification Challenge Dataset from 2018 to 2020. 16 , 24 This dataset contains 58,459 images of 9 skin cancer diseases: actinic keratosis, basal cell carcinoma, benign keratosis, dermatofibroma, melanoma, melanocytic nevus, squamous cell carcinoma, vascular lesion, and other unknown skin cancer cases.…”
Section: Methodsmentioning
confidence: 99%
“… 3 Combining and then jointly training a small dataset of interest with a larger auxiliary dataset often helps the prediction accuracy. 22 , 23 For the auxiliary dataset, we downloaded the publicly available SIIM-ISIC Melanoma Classification Challenge Dataset from 2018 to 2020. 16 , 24 This dataset contains 58,459 images of 9 skin cancer diseases: actinic keratosis, basal cell carcinoma, benign keratosis, dermatofibroma, melanoma, melanocytic nevus, squamous cell carcinoma, vascular lesion, and other unknown skin cancer cases.…”
Section: Methodsmentioning
confidence: 99%
“…Winata et al (2018) weighted losses for language modeling and POS tagging in an MTL setting, finding a lower weight to language modeling yielded a reduction in perplexity in modeling codeswitching between Chinese and English. A multitask supervised pretraining adaption strategy using a hierarchical architecture that learns multiple tasks on a source domain before fine-tuning them on the target was implemented by Meftah et al (2020). By using different weights for the different level tasks, starting with higher weights for lower tasks before incrementally increasing weights to higher level tasks during training, they achieve a noticeable error reduction in POS tagging, dependency parsing, and chunking.…”
Section: Mtl Performancementioning
confidence: 99%
“…MuTSPad (Multi-Task Supervised Pretraining and Adaptation) (Meftah et al, 2020) leverages hi-erarchical learning of a multi-task model on highresource domain followed by fine-tuning on multiple tasks on the low-resource target domain.…”
Section: Domainmentioning
confidence: 99%