Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.226
|View full text |Cite
|
Sign up to set email alerts
|

Domain Adaptation for Arabic Cross-Domain and Cross-Dialect Sentiment Analysis from Contextualized Word Embedding

Abstract: Finetuning deep pre-trained language models has shown state-of-the-art performances on a wide range of Natural Language Processing (NLP) applications. Nevertheless, their generalization performance drops under domain shift. In the case of Arabic language, diglossia makes building and annotating corpora for each dialect and/or domain a more challenging task. Unsupervised Domain Adaptation tackles this issue by transferring the learned knowledge from labeled source domain data to unlabeled target domain data. In… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
5

Relationship

3
7

Authors

Journals

citations
Cited by 20 publications
(10 citation statements)
references
References 36 publications
0
6
0
Order By: Relevance
“…For domain-specific data, domain adaptive finetuning of existing PLMs using MLM or domain adaptation have been shown to improve the performance of NLP applications (Rietzler et al, 2020;Barbieri et al, 2021;El Mekki et al, 2021a). Nevertheless, when the domain-specific data is sufficiently large, these transformers can be trained from scratch (Abdul-Mageed et al, 2021;Inoue et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…For domain-specific data, domain adaptive finetuning of existing PLMs using MLM or domain adaptation have been shown to improve the performance of NLP applications (Rietzler et al, 2020;Barbieri et al, 2021;El Mekki et al, 2021a). Nevertheless, when the domain-specific data is sufficiently large, these transformers can be trained from scratch (Abdul-Mageed et al, 2021;Inoue et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…In this work published in 2021, the authors proposed an unsupervised approach domain adaptation for Arabic Cross-Dialect and Cross-Domain based on the Word Embedding technique [10]. During the experimental phase, they adopted the fine-grained and coarse-grained taxonomies of Arabic dialects.…”
Section: Word Embeddingmentioning
confidence: 99%
“…To take profit from the provided unlabeled dataset in this shared task, we generate a weakly-annotated dataset and re-train the developed model on it. This method has been applied differently in several works (Khalifa et al, 2021;El Mekki et al, 2021a;Huang et al, 2021). In our work, we apply the following pipeline:…”
Section: Self-trainingmentioning
confidence: 99%