Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.134
|View full text |Cite
|
Sign up to set email alerts
|

Inexpensive Domain Adaptation of Pretrained Language Models: Case Studies on Biomedical NER and Covid-19 QA

Abstract: Domain adaptation of Pretrained LanguageModels (PTLMs) is typically achieved by unsupervised pretraining on target-domain text. While successful, this approach is expensive in terms of hardware, runtime and CO 2 emissions. Here, we propose a cheaper alternative: We train Word2Vec on target-domain text and align the resulting word vectors with the wordpiece vectors of a general-domain PTLM. We evaluate on eight English biomedical Named Entity Recognition (NER) tasks and compare against the recently proposed Bio… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
20
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2
2

Relationship

0
10

Authors

Journals

citations
Cited by 34 publications
(27 citation statements)
references
References 20 publications
0
20
0
Order By: Relevance
“…Nevertheless, an opposite way has also been observed, with authors searching for a green research that does not use models which are not environmentally friendly (expensive in terms of hardware, running time, and CO 2 footprint). This way has been investigated by Poerner et al [ 18 ] which proposed a GreenBioBERT model that has been produced using Word2vec to train a model on a new target domain (namely, on the COVID-19 issue) along with an alignment of vectors from the existing BioBERT model and the model trained with Word2vec.…”
Section: Current Trends In Biomedical Nlpmentioning
confidence: 99%
“…Nevertheless, an opposite way has also been observed, with authors searching for a green research that does not use models which are not environmentally friendly (expensive in terms of hardware, running time, and CO 2 footprint). This way has been investigated by Poerner et al [ 18 ] which proposed a GreenBioBERT model that has been produced using Word2vec to train a model on a new target domain (namely, on the COVID-19 issue) along with an alignment of vectors from the existing BioBERT model and the model trained with Word2vec.…”
Section: Current Trends In Biomedical Nlpmentioning
confidence: 99%
“…Among the three best papers selected by the Natural Language Processing (NLP) section, the paper by Poerner et al . presented a new energy-efficient transformer model, applied in particular to perform question-answering about COVID-19 [ 12 ]. As underlined by Natalia Grabar and Cyril Grouin, the co-editors of the NLP section, much work has been done this year in the NLP field on COVID-19, including the development of a dedicated corpus and the use of patient data, scientific literature, and social networks to predict or analyse COVID-19-related events [ 13 ].…”
Section: Highlights Of the 30th Edition Of The Imia Yearbookmentioning
confidence: 99%
“…Figure 3 shows that the gap between monolingual and multilingual tokenization quality is indeed larger in the specific texts (green bars) compared to the general texts (brown bars), indicating that in a specific domain, it is even harder for a multilingual model to outperform a monolingual model. This suggests that methods for explicitly adding representations of domain-specific words (Poerner et al, 2020;Schick and Schütze, 2020) could be a promising direction for improving our approach. Error analysis on financial sentence classification To provide a better insight into the difference between the mono and multi models, we compare the error predictions on the Danish FINNEWS dataset, since results in Table 4 show that the mono outperforms all multi models with a large margin on this dataset.…”
Section: Domain-specific Multilingual Representationsmentioning
confidence: 99%