2022
DOI: 10.48550/arxiv.2204.10290
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning to Revise References for Faithful Summarization

Abstract: In many real-world scenarios with naturally occurring datasets, reference summaries are noisy and contain information that cannot be inferred from the source text. On large news corpora, removing low quality samples has been shown to reduce model hallucinations. Yet, this method is largely untested for smaller, noisier corpora. To improve reference quality while retaining all data, we propose a new approach: to revise-not remove-unsupported reference content. Without ground-truth supervision, we construct synt… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 49 publications
(90 reference statements)
0
1
0
Order By: Relevance
“…This issue is particularly pronounced in tasks that require factual accuracy or domain-specific knowledge, where LLMs have been shown to fabricate details or make assertions without factual basis [14], [15], [16]. Moreover, the research has looked into the impact of training methodologies on the propensity of LLMs to hallucinate, and revealed that the choice of training datasets, the diversity of content, and the methods of data pre-processing play crucial roles in shaping the models' tendency to generate spurious outputs [17], [18], [19]. In addition to training data, the architectural aspects of LLMs have also been scrutinized, as certain designs may be more susceptible to hallucinations, suggesting a need for architectural innovations to mitigate this issue [10], [20].…”
Section: A Exploring the Phenomenon Of Llm Hallucinationsmentioning
confidence: 99%
“…This issue is particularly pronounced in tasks that require factual accuracy or domain-specific knowledge, where LLMs have been shown to fabricate details or make assertions without factual basis [14], [15], [16]. Moreover, the research has looked into the impact of training methodologies on the propensity of LLMs to hallucinate, and revealed that the choice of training datasets, the diversity of content, and the methods of data pre-processing play crucial roles in shaping the models' tendency to generate spurious outputs [17], [18], [19]. In addition to training data, the architectural aspects of LLMs have also been scrutinized, as certain designs may be more susceptible to hallucinations, suggesting a need for architectural innovations to mitigate this issue [10], [20].…”
Section: A Exploring the Phenomenon Of Llm Hallucinationsmentioning
confidence: 99%