Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.51
|View full text |Cite
|
Sign up to set email alerts
|

Arabizi Language Models for Sentiment Analysis

Abstract: Arabizi is a written form of spoken Arabic, relying on Latin characters and digits. It is informal and does not follow any conventional rules, raising many NLP challenges. In particular, Arabizi has recently emerged as the Arabic language in online social networks, becoming of great interest for opinion mining and sentiment analysis. Unfortunately, only few Arabizi resources exist and state-of-the-art language models such as BERT do not consider Arabizi.In this work, we construct and release two datasets: (i) … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 30 publications
(24 reference statements)
0
2
0
Order By: Relevance
“…Regarding the specific case of Arabic dialects written in Arabizi, a recent BERT-based model have been pretrained on 7 millions Egyptian tweets and displayed effective results on a sentiment analysis task (Baert et al, 2020). Another very recent model, at the date of writing, was pre-trained on 4 millions Algerian tweets and also demonstrated interesting results on sentiment analysis (Abdaoui et al, 2021).…”
Section: Discussionmentioning
confidence: 99%
“…Regarding the specific case of Arabic dialects written in Arabizi, a recent BERT-based model have been pretrained on 7 millions Egyptian tweets and displayed effective results on a sentiment analysis task (Baert et al, 2020). Another very recent model, at the date of writing, was pre-trained on 4 millions Algerian tweets and also demonstrated interesting results on sentiment analysis (Abdaoui et al, 2021).…”
Section: Discussionmentioning
confidence: 99%
“…Pre-trained language models (PLMs) have significantly advanced state-of-the-art on various natural language processing tasks, such as sentiment analysis (Bataa and Wu, 2019;Baert et al, 2020), text classification (Sun et al, 2019a;Arslan et al, 2021), and question answering (Yang et al, 2019). Despite the remarkable results, PLMs have a large number of parameters which make them expensive for deployment (Yang et al, 2019).…”
Section: Introductionmentioning
confidence: 99%