2020 IEEE 32nd International Conference on Tools With Artificial Intelligence (ICTAI) 2020
DOI: 10.1109/ictai50040.2020.00067
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Lingual Transfer Learning for Complex Word Identification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 27 publications
0
8
0
Order By: Relevance
“…Zaharia et al [189] experimented with several transformer-based models, such as Multilingual BERT (mBERT) [126] and XLM-RoBERTa [41], for cross-lingual CWI. Both mBERT and XLM-RoBERTa are multilingual masked language models that are pretrained on numerous languages.…”
Section: Lexical Complexity Prediction In Languages Other Than Englishmentioning
confidence: 99%
See 1 more Smart Citation
“…Zaharia et al [189] experimented with several transformer-based models, such as Multilingual BERT (mBERT) [126] and XLM-RoBERTa [41], for cross-lingual CWI. Both mBERT and XLM-RoBERTa are multilingual masked language models that are pretrained on numerous languages.…”
Section: Lexical Complexity Prediction In Languages Other Than Englishmentioning
confidence: 99%
“…XLM-RoBERTa is also pretrained on 100 languages, yet with more data [41]. Zaharia et al [189] tested these models performance on the WikiNews datasets provided by CWI-2018 [185]. They found that XLM-RoBERTa was the best performing model.…”
Section: Lexical Complexity Prediction In Languages Other Than Englishmentioning
confidence: 99%
“…The Oracle functions best when applied to multiple solutions, by jointly using them to obtain a final prediction. At the same time, Zaharia et al (2020) explored the power of Transformer-based models (Vaswani et al, 2017) in cross-lingual environments by using different training scenarios, depending on the scarcity of the resources: zero-shot, one-shot, as well as few-shot learning. Moreover, CWI can be also approached as a probabilistic task.…”
Section: Newsmentioning
confidence: 99%
“…Their approach based on the user's native language identifies complex terms by automatically detecting cognates and false friends, using distributional similarity computed from fastText (Bojanowski 2017: 135-146) word embeddings. Similar types of features are used in (Zaharia 2020). To calculate similarity measures between words, the authors apply a technique presented in (Conneau 2017) to learn a linear mapping of two vector spaces that represent monolingual fastText word embeddings (e.g., between Spanish and German) into the same vector space.…”
Section: Related Workmentioning
confidence: 99%