2019 22nd International Conference on Computer and Information Technology (ICCIT) 2019
DOI: 10.1109/iccit48885.2019.9038560
|View full text |Cite
|
Sign up to set email alerts
|

Authorship Attribution in Bangla literature using Character-level CNN

Abstract: Characters are the smallest unit of text that can extract stylometric signals to determine the author of a text. In this paper, we investigate the effectiveness of character-level signals in Authorship Attribution of Bangla Literature and show that the results are promising but improvable. The time and memory efficiency of the proposed model is much higher than the word level counterparts but accuracy is 2-5% less than the best performing word-level models. Comparison of various word-based models is performed … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 13 publications
(19 citation statements)
references
References 31 publications
0
18
0
Order By: Relevance
“…They built a corpus consisting of 3125 passages and gained the highest accuracy (96%) with random forest than Naive Bayes (62%) and decision tree (85%) classifiers. Khatun et al [6] introduced a character-level CNN for attributing Bengali authorship. This system's performance decreased with an increased number of authors and sample texts.…”
Section: B Bengali Language Based Authorship Classificationmentioning
confidence: 99%
See 4 more Smart Citations
“…They built a corpus consisting of 3125 passages and gained the highest accuracy (96%) with random forest than Naive Bayes (62%) and decision tree (85%) classifiers. Khatun et al [6] introduced a character-level CNN for attributing Bengali authorship. This system's performance decreased with an increased number of authors and sample texts.…”
Section: B Bengali Language Based Authorship Classificationmentioning
confidence: 99%
“…The layer-wise weight values are stored through the metafile. To investigate the effect of authorship classification performance, LSTM [19], Char-level-CNN [6], SVM [10], SGD [50], Multilingual pre-trained BERT (M-BERT) [51] and Distil-BERT [52] classifiers are also implemented on the same datasets.…”
Section: Ex(authormentioning
confidence: 99%
See 3 more Smart Citations