2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT) 2019
DOI: 10.1109/icecct.2019.8869504
|View full text |Cite
|
Sign up to set email alerts
|

Anempirical Analysis of Classification Models for Detection of Fake News Articles

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 2 publications
0
5
0
Order By: Relevance
“…n-grams, tf-idf and Gradient Boosting Classifier, were conducted by Wynne and Wint [30]. Kaur et al and Telang et al evaluated multiple combinations of various classifier and extraction techniques in the fake news detection task [15,29]. Usage of the ensemble methods for disinformation detection was explored in [17].…”
Section: Related Workmentioning
confidence: 99%
“…n-grams, tf-idf and Gradient Boosting Classifier, were conducted by Wynne and Wint [30]. Kaur et al and Telang et al evaluated multiple combinations of various classifier and extraction techniques in the fake news detection task [15,29]. Usage of the ensemble methods for disinformation detection was explored in [17].…”
Section: Related Workmentioning
confidence: 99%
“…Tokenization aids in the splitting of phrases and stop words help in removing words that don't add much value and therefore makes the phrase or sentence more concise. It has also been noticed that neural networks trained with N-gram which assigns probabil-ities to sentences and sequences of words performed better in comparison to sequence vectors [2] [3]. Furthermore, Models that are trained on a dataset consisting of news content rather than headlines tend to achieve higher accuracy and require more computational time.…”
Section: A Literature Surveymentioning
confidence: 99%
“…As mentioned in Section 3.2.4, mainstream datasets are relatively small for such a machine learning problem. This could be a reason as to why many authors combined datasets, in order to boost the number of articles for model training and to protect against issues such as overfitting, poor generalisability as well as bias (as justified in [61,62]). Existing annotation approaches, manual vs. automated annotation, have limitations which result in a trade-off; that is, datasets are: either well-labelled but smaller and expensive in terms of human labour, by using independent fact-checkers, or poorly labelled but larger using generic labelling algorithms, such as labelling based on the source domain name of the articles.…”
Section: Comparison Of Average Performance By Feature Group (Rq21)mentioning
confidence: 99%