Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.165
|View full text |Cite
|
Sign up to set email alerts
|

COVID-Fact: Fact Extraction and Verification of Real-World Claims on COVID-19 Pandemic

Abstract: We introduce a FEVER-like dataset COVID-Fact of 4, 086 claims concerning the COVID-19 pandemic. The dataset contains claims, evidence for the claims, and contradictory claims refuted by the evidence. Unlike previous approaches, we automatically detect true claims and their source articles and then generate counter-claims using automatic methods rather than employing human annotators. Along with our constructed resource, we formally present the task of identifying relevant evidence for the claims and verifying … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
52
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 37 publications
(53 citation statements)
references
References 26 publications
1
52
0
Order By: Relevance
“…Then, these systems evaluate whether the retrieved evidence sentences validate or contradict the claim, or whether there is not enough information to make a judgment. More recently, the SCIFACT (Wadden et al 2020) and COVIDFACT (Saakyan, Chakrabarty, and Muresan 2021) benchmarks re-purposed this framework for the sci-entific domain by releasing datasets of medical claims to be verified against scientific content (Wang et al 2020). While this framework has led to impressive advances in fact verification performance (Ye et al 2020;Pradeep et al 2021), current benchmarks assume that the available evidence database contains only valid, factual information.…”
Section: Claim Verificationmentioning
confidence: 99%
See 2 more Smart Citations
“…Then, these systems evaluate whether the retrieved evidence sentences validate or contradict the claim, or whether there is not enough information to make a judgment. More recently, the SCIFACT (Wadden et al 2020) and COVIDFACT (Saakyan, Chakrabarty, and Muresan 2021) benchmarks re-purposed this framework for the sci-entific domain by releasing datasets of medical claims to be verified against scientific content (Wang et al 2020). While this framework has led to impressive advances in fact verification performance (Ye et al 2020;Pradeep et al 2021), current benchmarks assume that the available evidence database contains only valid, factual information.…”
Section: Claim Verificationmentioning
confidence: 99%
“…During document retrieval, documents in the evidence repository that are relevant to the claim are selected. Existing methods typically use information retrieval methods to rank documents based on relevance Wadden et al 2020) or use public APIs of commercial document indices (Hanselowski et al 2019;Saakyan, Chakrabarty, and Muresan 2021) to crawl related documents. In the sentence retrieval stage, individual sentences from these retrieved documents are selected with respect to their relevance to the claim, often using textual entailment (Hanselowski et al 2019), or sentence similarity methods.…”
Section: Evidence Retrievalmentioning
confidence: 99%
See 1 more Smart Citation
“…Naturally, search based on ambiguous claims can yield poor quality search results, and thus insufficient evidence if included in a dataset to facilitate research on evidence-based fact-checking. To the best of our knowledge, this has not been considered with most existing datasets [28,1,19,23,3], though recent work on fact checking related to COVID-19 did usefully evaluate pipeline systems using Google as a baseline engine [21].…”
Section: Claim Ambiguitymentioning
confidence: 99%
“…The rise of misinformation has also prompted a great body of work, especially in natural language processing (NLP), on the automatic fact checking of claims [24,20,3,11,26,10,21]. Despite tremendous progress, however, the task remains quite challenging.…”
Section: Introductionmentioning
confidence: 99%