Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.750
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating the Factual Consistency of Abstractive Text Summarization

Abstract: The most common metrics for assessing summarization algorithms do not account for whether summaries are factually consistent with source documents. We propose a weakly-supervised, model-based approach for verifying factual consistency and identifying conflicts between source documents and generated summaries. Training data is generated by applying a series of rule-based transformations to the sentences of source documents. The factual consistency model is then trained jointly for three tasks: 1) predict whethe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

5
484
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 381 publications
(605 citation statements)
references
References 38 publications
5
484
0
Order By: Relevance
“…Various approaches have been proposed for text data augmentation, targeting improved performance on some diverse natural language processing (NLP) applications. Synthetic data can be generated using very simple rule-based transformations including noise injection (inserting random words), random word deletion or number swapping [21,22]. Another approach to creating synthetic text data are to randomly split training documents or sentences into multiple training fragments.…”
Section: Text Data Synthesismentioning
confidence: 99%
“…Various approaches have been proposed for text data augmentation, targeting improved performance on some diverse natural language processing (NLP) applications. Synthetic data can be generated using very simple rule-based transformations including noise injection (inserting random words), random word deletion or number swapping [21,22]. Another approach to creating synthetic text data are to randomly split training documents or sentences into multiple training fragments.…”
Section: Text Data Synthesismentioning
confidence: 99%
“…Another line of research focused on evaluating factual consistency of summarization systems. Kryscinski et al (2019b) proposed a weaklysupervised, model-based approach for evaluating factual consistency between source documents and generated summaries. They first generate training data by applying a series of transformations to randomly selected individual sentences from source documents (which they call claims) and assign them a binary label based on the type of the transformation.…”
Section: Related Workmentioning
confidence: 99%
“…Similarly, the authors of Falke et al (2019) evaluate three different state-of-the-art systems and find that between 8 and 26% of the generated summaries contain at least one factual error, even though ROUGE scores indicate good performance. Kryściński et al (2019) propose a weakly supervised method for verifying factual consistency between document and summary by training a binary model that predicts whether or not a sentence is consistent. For this purpose they artificially generate a dataset with various types of errors, such as entity or number swapping, paraphrasing, pronoun swapping, sentence negation and noise injection.…”
Section: Factual Errors In Summariesmentioning
confidence: 99%