2021
DOI: 10.3390/app11062567
|View full text |Cite
|
Sign up to set email alerts
|

Arabic Gloss WSD Using BERT

Abstract: Word Sense Disambiguation (WSD) aims to predict the correct sense of a word given its context. This problem is of extreme importance in Arabic, as written words can be highly ambiguous; 43% of diacritized words have multiple interpretations and the percentage increases to 72% for non-diacritized words. Nevertheless, most Arabic written text does not have diacritical marks. Gloss-based WSD methods measure the semantic similarity or the overlap between the context of a target word that needs to be disambiguated … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 18 publications
(9 citation statements)
references
References 20 publications
(30 reference statements)
0
9
0
Order By: Relevance
“…In this paper, we investigate the use of different types of signals to emphasize target words in context for Arabic WSD. El-Razzaz et al (2021) fine-tuned two BERT models on a small dataset of context-gloss pairs, consisting of about 5k lemmas, about 15k positive and 15k negative context-gloss pairs. They claimed an F1-score of 89%.…”
Section: Related Workmentioning
confidence: 99%
“…In this paper, we investigate the use of different types of signals to emphasize target words in context for Arabic WSD. El-Razzaz et al (2021) fine-tuned two BERT models on a small dataset of context-gloss pairs, consisting of about 5k lemmas, about 15k positive and 15k negative context-gloss pairs. They claimed an F1-score of 89%.…”
Section: Related Workmentioning
confidence: 99%
“…On the other hand, FP (false positive) count represents the number of class 2 samples that have been miss-classified as class 1 and vice-versa with FN (false negative) count. A specific part in the work done by El-Razzaz explains these metrics briefly and utilizes it in other evaluations [29].…”
Section: Resultsmentioning
confidence: 99%
“…TSV has proven to be an effective solution for the WSD in many state-of-the-art efforts. Although some researchers did not use the term TSV, this notion was also referred to as GlossBERT or Context-Gloss Binary Classification (Al-Hajj and Jarrar, 2022;El-Razzaz et al, 2021). A TSV training dataset is typically a set of context-gloss pairs, each labeled with Positive or Negative.…”
Section: Related Workmentioning
confidence: 99%
“…It can be fine-tuned on domain/task-specific data (e.g., POS tagging, WSD, TSV, and WiC) to update its contextualized embeddings. The TSV task has been addressed by fine-tuning BERT on context-gloss pairs as a sentence pair binary classification problem (Huang et al, 2019;Yap et al, 2020;Patel et al, 2021;Ranjbar and Zeinali, 2021;Lin and Giambi, 2021;El-Razzaz et al, 2021;Al-Hajj and Jarrar, 2022). However, the TSV task, similar to most NLP tasks, suffers from the knowledge-gain bottleneck, i.e., the lack of available quality datasets to train machine learning models.…”
Section: Introductionmentioning
confidence: 99%