2021
DOI: 10.1007/s11063-021-10528-4
|View full text |Cite
|
Sign up to set email alerts
|

ParsBERT: Transformer-based Model for Persian Language Understanding

Abstract: The surge of pre-trained language models has begun a new era in the field of Natural Language Processing (NLP) by allowing us to build powerful language models. Among these models, Transformer-based models such as BERT have become increasingly popular due to their stateof-the-art performance. However, these models are usually focused on English, leaving other languages to multilingual models with limited resources. This paper proposes a monolingual BERT for the Persian language (ParsBERT), which shows its stat… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
49
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 104 publications
(65 citation statements)
references
References 21 publications
0
49
0
Order By: Relevance
“…We design and implement three versions of a deep-learning based QA system and deploy PersianQuAD as the training set of the QA systems. In line with the state-of-the-art research on QA tasks [36], we used three pre-trained language models in our QA systems: MBERT [37], ParsBERT [38] and ALBERT-FA [39]. MBERT (Multilingual Bidirectional Encoder Representations from Transformers) is a deep bidirectional language model developed by Google.…”
Section: A Methodsmentioning
confidence: 99%
“…We design and implement three versions of a deep-learning based QA system and deploy PersianQuAD as the training set of the QA systems. In line with the state-of-the-art research on QA tasks [36], we used three pre-trained language models in our QA systems: MBERT [37], ParsBERT [38] and ALBERT-FA [39]. MBERT (Multilingual Bidirectional Encoder Representations from Transformers) is a deep bidirectional language model developed by Google.…”
Section: A Methodsmentioning
confidence: 99%
“…As regards the text encoder, we carried out the experiments with three different multilingual models, i.e., the multilingual version of BERT (Devlin et al, 2019) (mBERT) and the base and large versions of XLM-RoBERTa (Conneau et al, 2020) (XLMR-base and XLMR-large, respec-tively). In the monolingual setting, we used the following language-specific models: BERT-de 8 , CamemBERT-large (Martin et al, 2020) 9 , BERTit 10 , and ParsBERT 11 (Farahani et al, 2020), respectively, for German, French, Italian, and Farsi. As for all the other languages covered by the Word-Net datasets, i.e., Bulgarian, Chinese, Croatian, Danish, Dutch, Estonian, Japanese and Korean, we used the pre-trained models made available by TurkuNLP.…”
Section: Methodsmentioning
confidence: 99%
“…We have evaluated PQuAD using two pre-trained transformer-based language models, namely ParsBERT (Farahani et al, 2021) and XLM-RoBERTa (Conneau et al, 2020), as well as BiDAF (Levy et al, 2017) which is an attention-based model proposed for MRC.…”
Section: Modelsmentioning
confidence: 99%