Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.292
|View full text |Cite
|
Sign up to set email alerts
|

RobBERT: a Dutch RoBERTa-based Language Model

Abstract: Pre-trained language models have been dominating the field of natural language processing in recent years, and have led to significant performance gains for various complex natural language tasks. One of the most prominent pre-trained language models is BERT, which was released as an English as well as a multilingual version. Although multilingual BERT performs well on many tasks, recent studies show that BERT models trained on a single language significantly outperform the multilingual version. Training a Dut… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
71
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 134 publications
(96 citation statements)
references
References 25 publications
(31 reference statements)
0
71
1
Order By: Relevance
“…Since the pre-publication of this work (Martin et al, 2019), many monolingual language models have appeared, e.g. (Le et al, 2019;Virtanen et al, 2019;Delobelle et al, 2020), for as much as 30 languages (Nozza et al, 2020). In almost all tested configurations they displayed better results than multilingual language models such as mBERT (Pires et al, 2019).…”
Section: How Much Data Do You Need?mentioning
confidence: 99%
“…Since the pre-publication of this work (Martin et al, 2019), many monolingual language models have appeared, e.g. (Le et al, 2019;Virtanen et al, 2019;Delobelle et al, 2020), for as much as 30 languages (Nozza et al, 2020). In almost all tested configurations they displayed better results than multilingual language models such as mBERT (Pires et al, 2019).…”
Section: How Much Data Do You Need?mentioning
confidence: 99%
“…Predictions for both tasks were made with BERTje and RobBERT (de Vries et al, 2019;Delobelle et al, 2020; the Dutch versions of BERT and RobBERTa) using HuggingFace 4.0.0 (Wolf et al, 2020). In an attempt to improve these models, the "tags" method described above was used, but with the "<met>" (onset) and "</met>" (offset) placeholders for generic features and the same more fine-grained placeholders as described above when using source domain features.…”
Section: Bertje and Robbertmentioning
confidence: 99%
“…Since the BERT models were found to be effective for a wide range of NLP tasks (Devlin et al, 2019), several efforts have been extended towards improving them by more efficient training strategies Yang et al, 2019b;Sanh et al, 2019;Lan et al, 2019), training them for different domains Lee et al, 2019a;Lee and Hsiang, 2019;Chalkidis et al, 2020;Gururangan et al, 2020) and languages (Devlin, 2018;de Vries et al, 2019;Le et al, 2020;Martin et al, 2020;Delobelle et al, 2020;Cañete et al, 2020). Within the clinical domain, different models include the BioBERT models pretrained on PubMed abstracts and PMC full-text articles (Lee et al, 2019a), SciBERT trained on scientific text , clinicalBERT models trained on patient notes from the MIMIC-III corpus (Johnson et al, 2016) (sometimes as a continuation of the BioBERT models) (Alsentzer et al, 2019), and BlueBERT models that also use Pubmed abstracts and MIMIC-III patient notes for training .…”
Section: Related Workmentioning
confidence: 99%