2021
DOI: 10.1007/978-3-030-88942-5_33
|View full text |Cite
|
Sign up to set email alerts
|

The Case for Latent Variable Vs Deep Learning Methods in Misinformation Detection: An Application to COVID-19

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 15 publications
0
1
0
Order By: Relevance
“…By developing human-based evaluation metrics, we will not only assess the document embedding space, but more importantly, we will be able to identify potential biases related to certain characteristics of the collected abstracts enabling us to correct our model before it is deployed at scale. In addition, comparing BERT with other popular latent variable methods as presented in [10], would be of high interest. Finally, in terms of a computational chemistry perspective, the development of validation techniques for the extracted document embeddings and how they can be used for the discovery of energetic materials and systems is a significant research direction that deserves further investigation.…”
Section: Discussionmentioning
confidence: 99%
“…By developing human-based evaluation metrics, we will not only assess the document embedding space, but more importantly, we will be able to identify potential biases related to certain characteristics of the collected abstracts enabling us to correct our model before it is deployed at scale. In addition, comparing BERT with other popular latent variable methods as presented in [10], would be of high interest. Finally, in terms of a computational chemistry perspective, the development of validation techniques for the extracted document embeddings and how they can be used for the discovery of energetic materials and systems is a significant research direction that deserves further investigation.…”
Section: Discussionmentioning
confidence: 99%
“…The set of unique n-grams, iterations over the set, and count of the number of times each n-gram appears in the text are extracted to compute the unigram, bi-gram, and tri-gram overlap in order to quantify repetitiveness properties. The overlap is then calculated as a ratio between the count and the total number of different n-grams [38]. Additionally, the frequency of the words in the data is compared to the top 5K and 10K words in each language [24], thus determining how closely the lexicon in the dataset matches that of everyday speech.…”
mentioning
confidence: 99%