Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1433
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Domain Adaptation of Contextualized Embeddings for Sequence Labeling

Abstract: Contextualized word embeddings such as ELMo and BERT provide a foundation for strong performance across a wide range of natural language processing tasks by pretraining on large corpora of unlabeled text. However, the applicability of this approach is unknown when the target domain varies substantially from the pretraining corpus. We are specifically interested in the scenario in which labeled data is available in only a canonical source domain such as newstext, and the target domain is distinct from both the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
137
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 143 publications
(152 citation statements)
references
References 33 publications
(38 reference statements)
3
137
0
Order By: Relevance
“…We also experimented with searching concepts in SNOMED-CT using the Meaning Cloud tool 12 , however it did not work well, as many concepts for the shared task were annotated based on their synonyms.…”
Section: Number Of Training Epochsmentioning
confidence: 99%
See 1 more Smart Citation
“…We also experimented with searching concepts in SNOMED-CT using the Meaning Cloud tool 12 , however it did not work well, as many concepts for the shared task were annotated based on their synonyms.…”
Section: Number Of Training Epochsmentioning
confidence: 99%
“…The embedding function is trained either from a language modeling perspective [10] or based on recovering masked parts of tokens [11]. The downstream tasks which incorporate these embeddings are considered to be learned in a semi-supervised manner because they benefit from large amounts of unlabeled data [12], [13].…”
Section: Introductionmentioning
confidence: 99%
“…One possible direction to improve pre-trained multilingual LMs for MRC questions in a target language is to apply unsupervised domain adaptation [44]. Hundreds of questions would not be enough for a multilingual LM to fully adapt to both the target language and the downstream task.…”
Section: Modelmentioning
confidence: 99%
“…However, the target domain usually differs from the pre-training corpus which may result in the unsatisfactory performance of the model on the downstream task. Recently, unsupervised domain adaptation of language models has shown quality improvement in a number of NLP tasks, including sequence labelling [9]. [10] study unsupervised domain adaptation of BERT in the limited labelled and unlabelled data scenarios.…”
Section: Related Workmentioning
confidence: 99%