2017
DOI: 10.1016/j.jksuci.2016.11.010
|View full text |Cite
|
Sign up to set email alerts
|

Enhancing Arabic stemming process using resources and benchmarking tools

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 23 publications
(14 citation statements)
references
References 12 publications
0
13
0
Order By: Relevance
“…In spite of the increasing use of stemming as a requirement or a pre-processing step in different NLP applications, there is no stemming algorithm that is 100% precise. To address this problem, dissimilar studies have been recently focussed on evaluating and comparing the performance of Arabic stemmers to provide users and researchers with answers about the most appropriate algorithm for their tasks [7,9,10]. Nevertheless, there are no definite answers to the effectiveness of stemming in stylometric authorship applications in Arabic.…”
Section: Research Questionmentioning
confidence: 99%
“…In spite of the increasing use of stemming as a requirement or a pre-processing step in different NLP applications, there is no stemming algorithm that is 100% precise. To address this problem, dissimilar studies have been recently focussed on evaluating and comparing the performance of Arabic stemmers to provide users and researchers with answers about the most appropriate algorithm for their tasks [7,9,10]. Nevertheless, there are no definite answers to the effectiveness of stemming in stylometric authorship applications in Arabic.…”
Section: Research Questionmentioning
confidence: 99%
“…For our experiment we contacted the authors of all the aforementioned stemmers to share the source code. Only three agreed and shared the source, [11], [17], [18] for which we are grateful. ARLSTem [19], source is also freely available, but as it is in python, we re-implemented it in java.…”
Section: Related Workmentioning
confidence: 94%
“…On the other hand, LSTM or GRU models performed well either with bidirectional or with attention mechanism. What surprised us was the degraded performance by CNN-LSTM and CNN-GRU on the SPA corpus when used with the light stemmer [18]. For CNN-LSTM, the performance dropped from mid 90's (using any of the other stemming algorithms) to 75.4%, and for CNN-GRU it is even worse where it drops to ≈ 52%.…”
Section: Experimenting With Different Stemming Algorithmsmentioning
confidence: 95%
See 2 more Smart Citations