2019
DOI: 10.48550/arxiv.1910.08684
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sticking to the Facts: Confident Decoding for Faithful Data-to-Text Generation

Abstract: Neural conditional text generation systems have achieved significant progress in recent years, showing the ability to produce highly fluent text. However, the inherent lack of controllability in these systems allows them to hallucinate factually incorrect phrases that are unfaithful to the source, making them often unsuitable for many real world systems that require high degrees of precision. In this work, we propose a novel confidence oriented decoder that assigns a confidence score to each target position. T… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
32
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(41 citation statements)
references
References 26 publications
0
32
0
Order By: Relevance
“…Hallucination. The injection of false information is a well-known phenomena in data-to-text generation (Wiseman et al, 2017;Tian et al, 2019;Dhingra et al, 2019;Parikh et al, 2020), and machine translation (Koehn and Knowles, 2017;, image captioning (Rohrbach et al, 2018), exposure bias (Wang and Sennrich, 2020), and question answering (Feng et al, 2018). More related to the task of dialogue systems, Dušek et al (2018Dušek et al ( , 2020 demonstrate that state-of-the-art natural language generation (NLG) models suffer from the hallucination problem.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Hallucination. The injection of false information is a well-known phenomena in data-to-text generation (Wiseman et al, 2017;Tian et al, 2019;Dhingra et al, 2019;Parikh et al, 2020), and machine translation (Koehn and Knowles, 2017;, image captioning (Rohrbach et al, 2018), exposure bias (Wang and Sennrich, 2020), and question answering (Feng et al, 2018). More related to the task of dialogue systems, Dušek et al (2018Dušek et al ( , 2020 demonstrate that state-of-the-art natural language generation (NLG) models suffer from the hallucination problem.…”
Section: Related Workmentioning
confidence: 99%
“…This suggests that this inherent lack of controllability may be remedied by leveraging external oracle knowledge. However, existing approaches to knowledge grounding often suffer from a source-reference divergence problem whereby the reference contains additional factual information and simply training on the reference is insufficient to guarantee faithfulness to the source (Wiseman et al, 2017;Parikh et al, 2020;Tian et al, 2019). Consequently, ensuring the faithfulness of knowledge grounded dialogue systems-via precise alignment of the source and reference-remains an open challenge.…”
Section: Introductionmentioning
confidence: 99%
“…To reduce hallucinations in the reference-based setting, researchers have applied iterative training (Nie et al, 2019), post editing (Dong et al, 2020), soft constraints, e.g. attention manipulation (Kiddon et al, 2016;Hua and Wang, 2019;Tian et al, 2019; or optimal transport (Wang et al, 2020b), and template/scaffold guided schema (Liu et al, 2017;Wiseman et al, 2018;Moryossef et al, 2019;Ye et al, 2020;Shen et al, 2020;Li and Rush, 2020;Balakrishnan et al, 2019;Liu et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Concept Expansion. With limited number of entities and concepts as input, generation systems are often incapable of producing long text with rich content, resulting in hallucination (Wiseman et al, 2017;Tian et al, 2019). Therefore, from the often-abstract core concepts, we aim to predict more specific concepts that are also relevant to the given title.…”
Section: Content Item Augmentationmentioning
confidence: 99%