Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.61
|View full text |Cite
|
Sign up to set email alerts
|

Read before Generate! Faithful Long Form Question Answering with Machine Reading

Abstract: Long-form question answering (LFQA) aims to generate a paragraph-length answer for a given question. While current work on LFQA using large pre-trained model for generation are effective at producing fluent and somewhat relevant content, one primary challenge lies in how to generate a faithful answer that has less hallucinated content. We propose a new end-to-end framework that jointly models answer generation and machine reading. The key idea is to augment the generation model with fine-grained, answer-relate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 15 publications
(15 citation statements)
references
References 14 publications
(16 reference statements)
0
15
0
Order By: Relevance
“…This shares motivation with a line of work studying query-focused summarization (Xu and Lapata, 2020). Concurrent to our work, Su et al (2022) studies improving faithfulness of long-form answer through predicting and focusing on salient information in retrieved evidence document. Lastly, our work build up on three datasets containing longform answers (Kwiatkowski et al, 2019;Fan et al, 2019;Nakano et al, 2021) and extends the analysis of long-form answers from earlier studies (Krishna et al, 2021).…”
Section: Related Workmentioning
confidence: 73%
“…This shares motivation with a line of work studying query-focused summarization (Xu and Lapata, 2020). Concurrent to our work, Su et al (2022) studies improving faithfulness of long-form answer through predicting and focusing on salient information in retrieved evidence document. Lastly, our work build up on three datasets containing longform answers (Kwiatkowski et al, 2019;Fan et al, 2019;Nakano et al, 2021) and extends the analysis of long-form answers from earlier studies (Krishna et al, 2021).…”
Section: Related Workmentioning
confidence: 73%
“…Most developed language models produced nonfactual information [23], [24], [25] and [26] similar to other LLMs [27], [28], [29] and [30], the ChatGPT tool hallucinated some facts. For example, in Figure 6 below, when the researchers asked the tool about the flag of Yemen, a part of the generated response was correct, while another part of the information was revealed to be incorrect when verifying with the source.…”
Section: Redundancy Authenticity and Relatedness Principlesmentioning
confidence: 99%
“…We choose BART-large (Lewis et al, 2020), a transformer-based (Vaswani et al, 2017) generative pre-trained language model, as our backbone model for the generator because of its remarkable performance on text summarization benchmarks. Following the idea of Fusion-in-Decoder (FiD) and its applications in generation tasks (Izacard and Grave, 2021;Su et al, 2022;Vig et al, 2022), we employ FiD-BART for encoding multiple segments independently in the encoder and fuse information from all segments in the decoder jointly through the encoder-decoder attention.…”
Section: Generatormentioning
confidence: 99%